Hadoop的伪分布式安装

===================主机环境配置开始===============================

step1:安装VMware Workstation Pro

step2:安装Ctenos7

step3:在VMware中更改虚拟机网络类型为NAT方式(虚拟交换机的ip可以从vmvare的edit-->vertual network editor看到)

step4:修改网卡信息

#找到对应网卡
cd /etc/sysconfig/network-scripts/ vi ifcfg-ens32
#修改为static
BOOTPROTO=static
#修改为yes
ONBOOT=yes
#追加以下信息(NAT设置中可以找到IP填入)
GATEWAY=192.168.163.2
IPADDR=192.168.163.128
NETMASK=255.255.255.0
DNS1=114.114.114.114
DNS2=8.8.8.8
ARPCHECK=no

step5:重启网络服务 service network restart (重启之后可以ping一下是否通)

step6:修改主机名 vi /etc/sysconfig/network 添加一列 HOSTNAME=cMaster 

step7:修改主机名与IP映射,在root身份下 vi /etc/hosts  添加一行 192.168.2.100    cMaster

step8:关闭防火墙;

#查看防火墙状态
service iptables status
#临时关闭防火墙
service iptables stop
#关闭防火墙自动启动
chkconfig iptables off

======================主机环境完成================================

======================JDK安装开始================================

step1:下载jdk

step2:解压

step3:配置环境变量 vi /etc/profile (配置完成后执行 source /etc/profile 使其生效)

#配置环境变量
export JAVA_HOME=/home/ws/hadoopApp/jdk1.8.0_191
export HADOOP_HOME=/home/ws/hadoopApp/hadoop-3.0.2
export CLASSPATH=.:${JAVA_HOME}/lib.dt.jar:${JAVA_HOME}/lib/tools.jar

=====================JDK安装完成================================

====================Hadoop安装配置开始==========================

step1:下载Hadoop

step2:解压

step3:配置hadoop-env.sh(其配置文件均在hadoop-3.0.2/etc/hadoop下)

#修改JAVA_HOME为自定义安装的路径
export JAVA_HOME=/home/ws/hadoopApp/jdk1.8.0_191

step4:配置 core-site.xml

#添加如下代码

<configuration>

 <property>
  <name>hadoop.tmp.dir</name>
  <value>/home/ws/hadoopApp/cloudData</value>
 </property>

 <property>
  <name>fs.defaultFS</name>
  <value>hdfs://cMaster:9000/</value>
 </property>

</configuration>

step5:配置hdfs-site.xml

#添加如下代码

<configuration>
<property>
 <name>dfs.replication</name>
 <value>1</value>
</property>

<property>
 <name>dfs.http.address</name>
 <value>192.168.163.128:50070</value>
</property>

</configuration>

step6:配置mapred-site.xml(若文件名为mapred-site.xml.template,则先修改  mv mapred-site.xml.template mapred-site.xml )

#添加如下代码
<configuration>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
</configuration>

step7:配置yarn-site.xml

#添加如下代码
<configuration>

<!-- Site specific YARN configuration properties -->
<property>
<name>yarn.resourcemanager.hostname</name>
<value>cMaster</value>
</property>

<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>

</configuration>

step8:配置Hadoop环境变量( vi /etc/profile 同java配置)

step9:创建公钥 / 私钥

#生成秘钥
ssh-keygen -t rsa
#创建authorized_keys文件并修改权限为600 cd .ssh touch authorized_keys chmod 600 authorized_keys
#将公钥追加到authorized_keys文件中去 cat id_rsa.pub >> authorized_keys
#尝试能否成功免密登录 ssh cMaster

step10:初始化(格式化)

 hadoop namenode -format 

step11:启动

 start-dfs.sh

start-yarn.sh 

step12:查看

 jps 

step13:访问web管理页面

 #输入IP地址+端口号(50070)如: 192.168.163.128:5007 

step14:window下修改主机和IP映射

文件地址: C:\Windows\System32\drivers\etc\hosts

====================Hadoop安装配置j结束==========================

====================MapReduce测试==============================

 

使用MapReduce自带的统计词汇程序进行测试

step1:在hdfs下建立一个文件夹用于装输入文本

hadoop fs -mkdir hdfs://cmaster:9000/wordcount/input

step2:将测试文件上传

 hadoop fs -put test.txt /wordcount/input

step3:找到测试jar包

#hadoop-mapreduce-examples-3.0.2.jar测试jar包所在路径
cd /home/ws/hadoopApp/hadoop-3.0.2/share/hadoop/mapreduce/

step4:测试并将结果输入到指定文件夹

hadoop jar hadoop-mapreduce-examples-3.0.2.jar wordcount /wordcount/input /wordcount/output

注:测试时若报下列错误

[2018-11-18 20:57:15.662]Container exited with a non-zero exit code 1. Error file: prelaunch.err.
Last 4096 bytes of prelaunch.err :
Last 4096 bytes of stderr :
错误: 找不到或无法加载主类 org.apache.hadoop.mapreduce.v2.app.MRAppMaster

解决办法:

①输入  hadoop classpath 得到classpath

②在yarn-site.xml 中添加下列代码其中value就是classpath输出结果

<property>
<name>yarn.application.classpath</name>
<value>/home/ws/hadoopApp/hadoop-3.0.2/etc/hadoop:/home/ws/hadoopApp/hadoop-3.0.2/share/hadoop/common/lib/*:/home/ws/hadoopApp/hadoop-3.0.2/share/hadoop/common/*:/home/ws/hadoopApp/hadoop-3.0.2/share/hadoop/hdfs:/home/ws/hadoopApp/hadoop-3.0.2/share/hadoop/hdfs/lib/*:/home/ws/hadoopApp/hadoop-3.0.2/share/hadoop/hdfs/*:/home/ws/hadoopApp/hadoop-3.0.2/share/hadoop/mapreduce/*:/home/ws/hadoopApp/hadoop-3.0.2/share/hadoop/yarn:/home/ws/hadoopApp/hadoop-3.0.2/share/hadoop/yarn/lib/*:/home/ws/hadoopApp/hadoop-3.0.2/share/hadoop/yarn/*</value>
</property>

 

step5:查看结果

#查看目录
 hadoop fs -ls /wordcount/output
#查看结果文件
 hadoop fs -cat /wordcount/output/part-r-00000

或者在web端查看。

====================MapReduce测试结束==============================

 几篇安装参考网页

https://blog.csdn.net/hliq5399/article/details/78193113

https://blog.csdn.net/boom_man/article/details/78192385

https://www.cnblogs.com/thousfeet/p/8618696.html

https://www.cnblogs.com/zhangyinhua/p/7647686.html#_lab2_0_1

posted @ 2018-11-18 19:51  王小懒  阅读(269)  评论(0)    收藏  举报