https://www.zybuluo.com/DFFuture/note/626899
1.查看 hostname
修改主機名 sudo gedit /etc/hostname
重啟 reboot
2.編輯host sudo gedit /etc/hosts
刪除127.0.1.1那一行偏形,并添加:
192.168.86.140 ubuntu1
192.168.86.143 ubuntu2
192.168.86.144 ubuntu3
ping hadoop01捣炬、hadoop02、hadoop03看是否能
3.安裝openssh-server sudo apt-get install -y openssh-server
4.生成ssh目錄 ssh localhost
在ssh目錄下 cd ~./ssh
生成密鑰 ssh-keygen -t rsa -P ""
一直按回車即可
將公鑰添加到autorized_keys cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys
5.將公鑰傳輸?shù)狡渌?jié)點
scp ~/.ssh/id_rsa.pub edward@hadoop02:~
scp ~/.ssh/id_rsa.pub edward@hadoop03:~
進入hadoop02鼓黔、hadoop03目錄
將id_rea.pub的內(nèi)容添加進入authorized_keys
cat ~/id_rsa.pub >> ~/.ssh/authorized_keys
刪除已經(jīng)添加的id_rsa.pub
rm -rf id_rsa.pub
如果ssh hadoop01辫继、ssh hadoop02、ssh hadoop03都能無密登陸,即配置ssh成功
無法傳輸問題:
修改/etc/ssh/ssh_config
在最后添加2行:
StrictHostKeyChecking no
UserKnownHostsFile /dev/null
6.下載 hadoop hadoop-2.7.7.tar.gz
解壓到 /usr/local/hadoop
7.配置環(huán)境變量 gedit ~/.bashrc
添加
HADOOP_HOME=/usr/local/hadoop/hadoop-2.7.7
PATH添加:
$HADOOP_HOME/bin:$HADOOP_HOME/sbin
source ~/.bashrc hadoop version
8.集群/分布式模式需要修改 /usr/local/hadoop/hadoop-2.7.7/etc/hadoop 中的5個配置文件纳本,
8.1 slaves (配置DataNode主機)
寫入 :
hadoop01
hadoop02
hadoop03
8.2 core-site.xml (基礎(chǔ)配置)
寫入:
<configuration>
<property>
<name>fs.defaultFS</name>
<value>hdfs://hadoop02:9000</value>
</property>
<property>
<name>hadoop.tmp.dir</name>
<value>/local/hadoop/tmp</value>
</property>
<property>
<name>fs.trash.interval</name>
<value>10080</value>
</property>
</configuration>
8.3 hdfs-site.xml (HDFS配置)
寫入:
<configuration>
<property>
<name>dfs.replication</name>
<value>3</value>
</property>
<property>
<name>dfs.permissions.enabled</name>
<value>false</value>
</property>
<property>
<name>dfs.namenode.http-address</name>
<value>hadoop02:50070</value>
</property>
<property>
<name>dfs.namenode.secondary.http-address</name>
<value>hadoop03:50090</value>
</property>
</configuration>
8.4 mapred-site.xml (MapReduce配置)
復(fù)制重命名mapred-site.xml.template cp mapred-site.xml.template mapred-site.xml
寫入:
<configuration>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
<property>
<name>mapreduce.jobhistory.address</name>
<value>hadoop02:10020</value>
</property>
<property>
<name>mapreduce.jobhistory.webapp.address</name>
<value>hadoop02:19888</value>
</property>
<property>
<name>mapreduce.job.ubertask.enable</name>
<value>true</value>
</property>
</configuration>
8.5 yarn-site.xml (Yarn配置)
寫入:
<configuration>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<property>
<name>yarn.resourcemanager.hostname</name>
<value>hadoop01</value>
</property>
<property>
<name>yarn.web-proxy.address</name>
<value>hadoop01:8888</value>
</property>
<property>
<name>yarn.log-aggregation-enable</name>
<value>true</value>
</property>
<property>
<name>yarn.log-aggregation.retain-seconds</name>
<value>604800</value>
</property>
</configuration>
9.格式化nameNode hadoop namenode -format
異常問題可以添加權(quán)限 sudo chmod 777 / usr /local/ hadoop /tmp/
10.將hadoop 復(fù)制到其他主機中
scp -r /local/hadoop hadoop@hadoop02:/local/
scp -r /local/hadoop hadoop@hadoop03:/local/
啟動
11.hadoop02 /sbin 啟動dfs和jobhistory
start-dfs.sh
mr-jobhistory-daemon.sh start historyserver
12.Hadoop01 /sbin 啟動yarn和防護進程
start-yarn.sh
yarn-daemon.sh start proxyserver
13.網(wǎng)址
Yarn:192.168.86.136:8088
HDFS:192.168.86.137:50070
關(guān)閉
14. hadoop01 /sbin 關(guān)閉yarn
stop-yarn.sh
15 hadoop02 /sbin 關(guān)閉hdfs和jobhistory
stop-dfs.sh
mr-jobhistory-daemon.sh stop historyserver
報錯:
JAVA_HOME is not set and could not be found
https://blog.csdn.net/lxpbs8851/article/details/45225171
$HADOOP_HOME/etc/hadoop/yarn-env.sh 中添加
export JAVA_HOME=/usr/lib/jvm/jdk1.8.0_191
$HADOOP_HOME/libexec/hadoop-config.sh 中添加
export JAVA_HOME=/usr/lib/jvm/jdk1.8.0_191