2018-09-25 17:12:54 409浏览
今天扣丁学堂大数据培训老师带你探索Hadoop分布式集群的详细介绍,首先Hadoop的搭建有三种方式,单机版适合开发调试;伪分布式版,适合模拟集群学习;完全分布式,生产使用的模式。这篇文件介绍如何搭建完全分布式的hadoop集群,一个主节点,三个数据节点为例来讲解。
192.168.10.131 192.168.10.132 192.168.10.133 192.168.10.134
vim /etc/hosts 192.168.10.131 master 192.168.10.132 slave1 192.168.10.133 slave2 192.168.10.134 slave3
vi /etc/sysconfig/network HOSTNAME=master
yum -y install java-1.8.0-openjdk*
export JAVA_HOME=/usr/lib/jvm/jre-1.8.0-openjdk.x86_64 export PATH=$JAVA_HOME/bin:$PATH export CLASSPATH=.:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar
source /etc/profile
service iptables status
service iptables stop chkconfig iptables off
# vim /etc/selinux/config
#SELINUX=enforcing #SELINUXTYPE=targeted
—添加
SELINUX=disabled
ssh-keygen -t rsa
cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys
chmod 600 .ssh/authorized_keys
ssh master
scp root@master:/root/.ssh/id_rsa.pub /root/
cat id_rsa.pub >> .ssh/authorized_keys rm -rf id_rsa.pub
ssh slave1
scp root@slave1:/root/.ssh/id_rsa.pub /root/
cat id_rsa.pub >> .ssh/authorized_keys rm -rf id_rsa.pub //删除id_rsa.pub
ssh master
wget http://apache.claz.org/hadoop/common/hadoop-2.7.3/hadoop-2.7.
tar -xzvf hadoop-2.7.3.tar.gz -C /usr/local
mv hadoop-2.7.3 hadoop
export HADOOP_HOME=/usr/local/hadoop export PATH=$PATH:$HADOOP_HOME/bin
source /etc/profile
<configuration> <property> <name>hadoop.tmp.dir</name> <value>file:/usr/local/hadoop/tmp</value> <description>Abase for other temporary directories.</description> </property> <property> <name>fs.defaultFS</name> <value>hdfs://master:9000</value> </property> </configuration>
<configuration> <property> <name>dfs.replication</name> <value>3</value> </property> <property> <name>dfs.name.dir</name> <value>/usr/local/hadoop/hdfs/name</value> </property> <property> <name>dfs.data.dir</name> <value>/usr/local/hadoop/hdfs/data</value> </property> </configuration>
cp /usr/local/hadoop/etc/hadoop/mapred-site.xml.template /usr/local/hadoop/etc/hadoop/mapred-site.xml vim /usr/local/hadoop/etc/hadoop/mapred-site.xml <configuration> <property> <name>mapreduce.framework.name</name> <value>yarn</value> </property> <property> <name>mapred.job.tracker</name> <value>http://master:9000</value> </property> </configuration>
<configuration> <!– Site specific YARN configuration properties –> <property> <name>yarn.nodemanager.aux-services</name> <value>mapreduce_shuffle</value> </property> <property> <name>yarn.resourcemanager.hostname</name> <value>master</value> </property> </configuration>
vi /usr/local/hadoop/etc/hadoop/masters
master
vi /usr/local/hadoop/etc/hadoop/slaves ## 内容 slave1 slave2 slave3
scp -r /usr/local/hadoop slave1:/usr/local/
rm -rf /usr/local/hadoop/etc/hadoop/slaves
vi /etc/profile ## 内容 export HADOOP_HOME=/usr/local/hadoop export PATH=$PATH:$HADOOP_HOME/bin 使得hadoop命令在当前终端立即生效; source /etc/profile 依次配置其它slave服务
bin/hadoop namenode -format
sbin/start-all.sh
#master 执行 jps查看运行情况 25928 SecondaryNameNode 25742 NameNode 26387 Jps 26078 ResourceManager #slave 执行 jps查看运行情况 24002 NodeManager 23899 DataNode 24179 Jps
hadoop dfsadmin -report
Configured Capacity: 50108030976 (46.67 GB) Present Capacity: 41877471232 (39.00 GB) DFS Remaining: 41877385216 (39.00 GB) DFS Used: 86016 (84 KB) DFS Used%: 0.00% Under replicated blocks: 0 Blocks with corrupt replicas: 0 Missing blocks: 0 Missing blocks (with replication factor 1): 0 ……
sbin/stop-all.sh sbin/start-all.sh
vi /usr/local/hadoop/etc/hadoop/hadoop-env.sh ## 配置项 export JAVA_HOME=/usr/lib/jvm/jre-1.7.0-openjdk.x86_64
— 关闭SELINUX # vim /etc/selinux/config — 注释掉 #SELINUX=enforcing #SELINUXTYPE=targeted — 添加 SELINUX=disabled
以上就是关于扣丁学堂大数据培训带你探索Hadoop分布式集群的详细介绍,希望对同学们有所帮助,扣丁学堂提供完整大数据学习路线内容包含Linux&&Hadoop生态体系、大数据计算框架体系、云计算体系、机器学习&&深度学习等大数据视频教程内容,根据大数据学习路线图可以让你对学习大数据需要掌握的知识有个清晰的了解,并快速入门大数据开发,大数据学习群:827371640
【关注微信公众号获取更多学习资料】