3.3、配置slaves文件-->>增加slave节点
hadoop01
hadoop02 hadoop03 3.4、配置core-site.xml文件-->>增加hadoop核心配置(hdfs文件端口号是9000 、file:/home/luffy/opt/hadoop-2.6-0/tmp、) <configuration> <property> <name>fs.defaultFS</name> <value>hdfs://hadoop01:9000</value> </property> <property> <name>io.file.buffer.size</name> <value>131072</value> </property> <property> <name>hadoop.tmp.dir</name> <value>file:/home/luffy/opt/hadoop-2.6.0/tmp</value> <description>Abasefor other temporary directories.</description> </property> <property> <name>hadoop.proxyuser.luffy.hosts</name> <value>*</value> </property> <property> <name>hadoop.proxyuser.luffy.groups</name> <value>*</value> </property> </configuration> 3.5、配置hdfs-site.xml文件-->>增加hdfs配置信息(namenode、datanode端口号和目录位置) <configuration> <property> <name>dfs.namenode.secondary.http-address</name> <value>hadoop01:9001</value> </property> <property> <name>dfs.namenode.name.dir</name> <value>file:/home/luffy/opt/hadoop-2.6.0/dfs/name</value> </property> <property> <name>dfs.datanode.data.dir</name> <value>file:/home/luffy/opt/hadoop-2.6.0/dfs/data</value> </property> <property> <name>dfs.replication</name> <value>3</value> </property> <property> <name>dfs.webhdfs.enabled</name> <value>true</value> </property> </configuration> 3.6、配置mapred-site.xml文件-->>增加mapreduce配置(使用yarn框架、jobhistory使用地址以及web地址) <configuration> <property> <name>mapreduce.framework.name</name> <value>yarn</value> </property> <property> <name>mapreduce.jobhistory.address</name> <value>hadoop01:10020</value> </property> <property> <name>mapreduce.jobhistory.webapp.address</name> <value>hadoop01:19888</value> </property> </configuration> 3.7、配置yarn-site.xml文件-->>增加yarn功能 <configuration> <property> <name>yarn.nodemanager.aux-services</name> <value>mapreduce_shuffle</value> </property> <property> <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name> <value>org.apache.hadoop.mapred.ShuffleHandler</value> </property> <property> <name>yarn.resourcemanager.address</name> <value>hadoop01:8032</value> </property> <property> <name>yarn.resourcemanager.scheduler.address</name> <value>hadoop01:8030</value> </property> <property> <name>yarn.resourcemanager.resource-tracker.address</name> <value>hadoop01:8035</value> </property> <property> <name>yarn.resourcemanager.admin.address</name> <value>hadoop01:8033</value> </property> <property> <name>yarn.resourcemanager.webapp.address</name> <value>hadoop01:8088</value> </property> </configuration> 4、将配置好的hadoop文件copy到另一台slave机器上(完成copy一定要检查环境变量的配置) scp -r /opt/hadoop-2.6.0/ luffy@hadoop02:~/opt/ 四、验证 1,格式化namenode: 在hadoop01中操作 进入 cd hadoop-2.6.0 命令:./bin/hdfs namenode -format 5、启动hdfs 命令:./sbin/start-dfs.sh 命令:jps 4947 NameNode 5438 Jps 5068 DataNode 5216 SecondaryNameNode 配置hadoop命令 vi ~/.bashrc export JAVA_HOME=/home/chao/jdk1.7.0_79 export HADOOP_HOME=/home/chao/hadoop-2.6.0 export CLASSPATH=${JAVA_HOME}/lib export PATH=.:$HADOOP_HOME/bin:${JAVA_HOME}/bin:$PATH hadoop搭建完成