虚拟机搭建hadoop2.7.4
2017-09-02 12:11
204 查看
CentOS7.0安装配置hadoop2.7.4
资源准备
资源下载:hadoop官网下载链接 至于安装什么版本,看自己需求jdk官网下载链接注意事项:注意hadoop,jdk,centos都应该是64位或者32位的,以免出现无法预料的错误建议使用64位的linux 虚拟机配置
系统配置:虚拟机:一个master(Master.Hadoop),两个slave(Slave1.Hadoop, Slave2.Hadoop)网络设置:我NAT的方式内存:每个虚拟机配置1024M内存分区:自动软件选择:最小安装,注意选择开发工具进行以下步骤前,确保3台虚拟机与主机之间可以相互ping通参考虚拟机无法ping通、无法联网进行设置yum search ifconfg yum install net-tools.x86_6412
![](http://static.blog.csdn.net/images/save_snippets.png)
主机 | ip地址 |
---|---|
master.hadoop | 192.168.202.128 |
slave1.hadoop | 192.168.202.129 |
slave2.hadoop | 192.168.202.130 |
vi /etc/hosts // 将以下数据复制进入各个主机中 192.168.202.128 master.hadoop 192.168.202.129 slave1.hadoop 192.168.202.130 slave2.hadoop1234567
![](http://static.blog.csdn.net/images/save_snippets.png)
ping slave1.hadoopping slave2.hadoop123
![](http://static.blog.csdn.net/images/save_snippets.png)
配置Master无密码登录所有Salve
以下在Master主机上配置输入以下指令生成ssh,过程中遇到需要输入密码,只需执行回车Enter即可ssh-keygen// 会生成两个文件,放到默认的/root/.ssh/文件夹中123
![](http://static.blog.csdn.net/images/save_snippets.png)
cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys1
![](http://static.blog.csdn.net/images/save_snippets.png)
chmod 600 ~/.ssh/authorized_keys1
![](http://static.blog.csdn.net/images/save_snippets_01.png)
vi /etc/ssh/sshd_config// 以下三项修改成以下配置RSAAuthentication yes # 启用 RSA 认证PubkeyAuthentication yes # 启用公钥私钥配对认证方式AuthorizedKeysFile .ssh/authorized_keys # 公钥文件路径(和上面生成的文件同)123456789
![](http://static.blog.csdn.net/images/save_snippets.png)
service sshd restart1
![](http://static.blog.csdn.net/images/save_snippets_01.png)
// scp ~/.ssh/id_rsa.pub 远程用户名@远程服务器IP:~/scp ~/.ssh/id_rsa.pub root@192.168.202.129:~/scp ~/.ssh/id_rsa.pub root@192.168.202.130:~/1234
![](http://static.blog.csdn.net/images/save_snippets.png)
mkdir ~/.ssh// 修改权限chmod 700 ~/.ssh1234
![](http://static.blog.csdn.net/images/save_snippets.png)
cat ~/id_rsa.pub >> ~/.ssh/authorized_keys// 修改权限chmod 600 ~/.ssh/authorized_keys1234
![](http://static.blog.csdn.net/images/save_snippets.png)
rm –r ~/id_rsa.pub1
![](http://static.blog.csdn.net/images/save_snippets_01.png)
ssh 192.168.202.129ssh 192.168.202.130// 如果能够分别无密码登陆slave1, slave2主机,则成功配置1234
![](http://static.blog.csdn.net/images/save_snippets.png)
进行jdk, hadoop软件安装
jdk安装:在/usr下创建java文件夹使用第三方软件将jdk压缩包文件传到3台虚拟主机中使用以下指令进行加压tar zxvf jdk-8u45-linux-x64.tar.gz// 解压后可以删除掉gz文件rm jdk-8u45-linux-x64.tar.gz1234
![](http://static.blog.csdn.net/images/save_snippets.png)
vi /etc/profile12
![](http://static.blog.csdn.net/images/save_snippets_01.png)
// 将以下数据复制到文件底部export JAVA_HOME=/usr/java/jdk1.8.0_45export JRE_HOME=/usr/java/jdk1.8.0_45/jreexport CLASSPATH=.:$CLASSPATH:$JAVA_HOME/lib:$JRE_HOME/libexport PATH=$PATH:$JAVA_HOME/bin:$JRE_HOME/bin12345678910
![](http://static.blog.csdn.net/images/save_snippets.png)
source /etc/profile1
![](http://static.blog.csdn.net/images/save_snippets_01.png)
java -version// 如果出现以下信息,则配置成功java version "1.8.0_45"Java(TM) SE Runtime Environment (build 1.8.0_45-b14)Java HotSpot(TM) 64-Bit Server VM (build 25.45-b02, mixed mode)123456
![](http://static.blog.csdn.net/images/save_snippets.png)
cd /usrtar zxvf hadoop-2.7.4.tar.gzmv hadoop-2.7.4 hadoop// 删除hadoop-2.7.4.tar.gz文件rm –rf hadoop-2.7.4.tar.gz123456
![](http://static.blog.csdn.net/images/save_snippets.png)
cd /usr/hadoopmkdir tmp12
![](http://static.blog.csdn.net/images/save_snippets.png)
vi /etc/profile// 将以下数据加入到文件末尾export HADOOP_INSTALL=/usr/hadoopexport PATH=${HADOOP_INSTALL}/bin:${HADOOP_INSTALL}/sbin:${PATH}export HADOOP_MAPRED_HOME=${HADOOP_INSTALL}export HADOOP_COMMON_HOME=${HADOOP_INSTALL}export HADOOP_HDFS_HOME=${HADOOP_INSTALL}export YARN_HOME=${HADOOP_INSTALLL}export HADOOP_COMMON_LIB_NATIVE_DIR=${HADOOP_INSTALL}/lib/natvieexport HADOOP_OPTS="-Djava.library.path=${HADOOP_INSTALL}/lib:${HADOOP_INSTALL}/lib/native"123456789101112
![](http://static.blog.csdn.net/images/save_snippets.png)
source /etc/profile1
![](http://static.blog.csdn.net/images/save_snippets.png)
配置hadoop(先只在Master主机配置,配置完后传入两个Slave主机)
设置hadoop-env.sh和yarn-env.sh中的java环境变量cd /usr/hadoop/etc/hadoop/vi hadoop-env.sh// 修改JAVA_HOMEexport JAVA_HOME=/usr/java/jdk1.8.0_4512345
![](http://static.blog.csdn.net/images/save_snippets.png)
vi core-site.xml// 修改文件内容为以下<configuration><property><name>hadoop.tmp.dir</name><value>/usr/hadoop/tmp</value><description>A base for other temporary directories.</description></property><property><name>fs.default.name</name><value>hdfs://master.hadoop:9000</value></property></config108fauration>1234567891011121314151617181920212223
![](http://static.blog.csdn.net/images/save_snippets.png)
vi hdfs-site.xml// 修改文件内容为以下<configuration><property><name>dfs.namenode.name.dir</name><value>file:///usr/hadoop/dfs/name</value></property><property><name>dfs.datanode.data.dir</name><value>file:///usr/hadoop/dfs/data</value></property><property><name>dfs.replication</name><value>1</value></property><property><name>dfs.nameservices</name><value>hadoop-cluster1</value></property><property><name>dfs.namenode.secondary.http-address</name><value>master.hadoop:50090</value></property><property><name>dfs.webhdfs.enabled</name><value>true</value></property></configuration>123456789101112131415161718192021222324252627282930313233
![](http://static.blog.csdn.net/images/save_snippets.png)
vi mapred-site.xml// 修改文件为以下<property><name>mapreduce.framework.name</name><value>yarn</value><final>true</final></property><property><name>mapreduce.jobtracker.http.address</name><value>master.hadoop:50030</value></property><property><name>mapreduce.jobhistory.address</name><value>master.hadoop:10020</value></property><property><name>mapreduce.jobhistory.webapp.address</name><value>master.hadoop:19888</value></property><property><name>mapred.job.tracker</name><value>http://master.hadoop:9001</value></property>1234567891011121314151617181920212223242526
![](http://static.blog.csdn.net/images/save_snippets.png)
vi yarn-site.xml// 修改文件内容为以下<property><name>yarn.resourcemanager.hostname</name><value>master.hadoop</value></property><property><name>yarn.nodemanager.aux-services</name><value>mapreduce_shuffle</value></property><property><name>yarn.resourcemanager.address</name><value>master.hadoop:8032</value></property><property><name>yarn.resourcemanager.scheduler.address</name><value>master.hadoop:8030</value></property><property><name>yarn.resourcemanager.resource-tracker.address</name><value>master.hadoop:8031</value></property><property><name>yarn.resourcemanager.admin.address</name><value>master.hadoop:8033</value></property><property><name>yarn.resourcemanager.webapp.address</name><value>master.hadoop:8088</value></property>12345678910111213141516171819202122232425262728293031323334
![](http://static.blog.csdn.net/images/save_snippets.png)
配置Hadoop的集群
将Master中配置好的hadoop传入两个Slave中scp -r /usr/hadoop root@192.168.202.129:/usr/scp -r /usr/hadoop root@192.168.202.130:/usr/12
![](http://static.blog.csdn.net/images/save_snippets.png)
cd /usr/hadoop/etc/hadoopvi slaves// 将文件内容修改为slave1.hadoopslave2.hadoop1234567
![](http://static.blog.csdn.net/images/save_snippets.png)
// 在Master主机上输入以下指令hadoop namenode -format12
![](http://static.blog.csdn.net/images/save_snippets.png)
// 关闭机器防火墙 根据自己的版本不同,命令有所不同CentOS 7关闭firewall:systemctl stop/start firewalld.service 停止/启动firewallsystemctl disable firewalld.service 禁止firewall开机启动firewall-cmd --state 查看默认防火墙状态cd /usr/hadoop/sbin./start-all.sh// 更推荐的运行方式:cd /usr/hadoop/sbin./start-dfs.sh./start-yarn.sh应该输出以下信息:Starting namenodes on [Master.Hadoop]Master.Hadoop: starting namenode, logging to /usr/hadoop/logs/hadoop-root-namenode-localhost.localdomain.outSlave2.Hadoop: starting datanode, logging to /usr/hadoop/logs/hadoop-root-datanode-Slave2.Hadoop.outSlave1.Hadoop: starting datanode, logging to /usr/hadoop/logs/hadoop-root-datanode-Slave1.Hadoop.outstarting yarn daemonsstarting resourcemanager, logging to /usr/hadoop/logs/yarn-root-resourcemanager-localhost.localdomain.outSlave1.Hadoop: starting nodemanager, logging to /usr/hadoop/logs/yarn-root-nodemanager-Slave1.Hadoop.outSlave2.Hadoop: starting nodemanager, logging to /usr/hadoop/logs/yarn-root-nodemanager-Slave2.Hadoop.out12345678910111213141516171819202122232425
![](http://static.blog.csdn.net/images/save_snippets.png)
// 1. 直接在Master或Slave输入指令:jps// 应该输出以下信息(端口号仅供参考)Master:3930 ResourceManager4506 Jps3693 NameNodeSlave:2792 NodeManager2920 Jps2701 DataNode// 2. 输入以下指令hadoop dfsadmin -report// 应该输出以下信息:Configured Capacity: 14382268416 (13.39 GB)Present Capacity: 10538565632 (9.81 GB)DFS Remaining: 10538557440 (9.81 GB)DFS Used: 8192 (8 KB)DFS Used%: 0.00%Under replicated blocks: 0Blocks with corrupt replicas: 0Missing blocks: 0Missing blocks (with replication factor 1): 0-------------------------------------------------Live datanodes (2):Name: 192.168.1.124:50010 (Slave2.Hadoop)Hostname: Slave2.HadoopDecommission Status : NormalConfigured Capacity: 7191134208 (6.70 GB)DFS Used: 4096 (4 KB)Non DFS Used: 1921933312 (1.79 GB)DFS Remaining: 5269196800 (4.91 GB)DFS Used%: 0.00%DFS Remaining%: 73.27%Configured Cache Capacity: 0 (0 B)Cache Used: 0 (0 B)Cache Remaining: 0 (0 B)Cache Used%: 100.00%Cache Remaining%: 0.00%Xceivers: 1Last contact: Thu Jul 02 10:45:04 CST 2015Name: 192.168.1.125:50010 (Slave1.Hadoop)Hostname: Slave1.HadoopDecommission Status : NormalConfigured Capacity: 7191134208 (6.70 GB)DFS Used: 4096 (4 KB)Non DFS Used: 1921769472 (1.79 GB)DFS Remaining: 5269360640 (4.91 GB)DFS Used%: 0.00%DFS Remaining%: 73.28%Configured Cache Capacity: 0 (0 B)Cache Used: 0 (0 B)Cache Remaining: 0 (0 B)Cache Used%: 100.00%Cache Remaining%: 0.00%Xceivers: 112345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364
![](http://static.blog.csdn.net/images/save_snippets.png)
// CentOS7中iptables这个服务的配置文件没有了,采用了新的firewalld// 输入以下指令后,可以在真机浏览器上访问hadoop网页systemctl stop firewalld// 输入以下网页,进入hadoop管理首页(IP地址为master 主机IP) http://192.168.202.128:50070/dfshealth.html#tab-overview1234567
![](http://static.blog.csdn.net/images/save_snippets.png)
相关文章推荐
- Windows 7平台利用Vmware Workstation 11虚拟机搭建Hadoop 2.7.4基于Ubuntu kylin 14.04集群环境
- 如何在虚拟机上搭建并配置一个分布式的 Hadoop2.7.4 集群
- (一)hadoop系列之__XP环境下搭建linux虚拟机
- spark+hadoop集群搭建-虚拟机
- Hadoop-2.7.4 集群快速搭建
- 在Win7虚拟机下搭建Hadoop2.5.2+Spark1.5.2单机环境
- 在Win7虚拟机下搭建Hadoop2.6.0伪分布式环境
- hadoop集群搭建(虚拟机)
- 联想ThinkPad S3-S440虚拟机安装,ubuntu安装,Hadoop(2.7.1)详解及WordCount运行,spark集群搭建
- 虚拟机搭建hadoop的全分布式集群-in detail (4)
- 搭建VM的虚拟机下hadoop集群环境
- 虚拟机中搭建hadoop
- java 虚拟机 hadoop 环境搭建 (本次搭建全过程)
- 轻松搭建hadoop1.2.1集群前奏--虚拟机安装与配置
- 虚拟机搭建hadoop集群
- (二)hadoop系列之__linux虚拟机搭建JDK和Eclipse环境
- hadoop集群搭建--于虚拟机中
- CentOS环境下模拟Hadoop2.7.3环境搭建(虚拟机)
- Windows下虚拟机搭建Hadoop集群
- 在hadoop2.7.4中搭建apache-storm-1.1.1集群