当前位置 博文首页 > Andywu2019的博客:Hadoop2.7.2伪分布式安装步骤

    Andywu2019的博客:Hadoop2.7.2伪分布式安装步骤

    作者:[db:作者] 时间:2021-09-21 11:58

    安装Hadoop的准备工作:

    1、修改/etc/hosts? 2、配置JDK? 3、关闭防火墙? 4、配置SSH(免密码登录)

    # tar zxf hadoop-2.7.2.tar.gz? (安装及配置Hadoop)

    修改配置文件:

    cd /opt/hadoop/hadoop-2.7.2/etc/hadoop/

    修改配置文件:

    1、core-site.xml

    <configuration>

    ?<property>

    ??? <name>fs.default.name</name>

    ??? <value>hdfs://centos:9000</value>

    ? </property>

    ?<property>

    ??? <name>hadoop.tmp.dir</name>

    ??? <value>/opt/hadoop/hadoop-2.7.2/current/tmp</value>

    ? </property>

    ?<property>

    ??? <name>fs.trash.interval</name>

    ??? <value>4320</value>

    ? </property>

    </configuration>

    2、hdfs-site.xml

    <configuration>

    ?<property>

    ?? <name>dfs.namenode.name.dir</name>

    ?? <value>/opt/hadoop/hadoop-2.7.2/current/dfs/name</value>

    ?</property>

    ?<property>

    ?? <name>dfs.datanode.data.dir</name>

    ?? <value>/opt/hadoop/hadoop-2.7.2/current/data</value>

    ?</property>

    ?<property>

    ?? <name>dfs.replication</name>

    ?? <value>1</value>

    ?</property>

    ?<property>

    ?? <name>dfs.webhdfs.enabled</name>

    ?? <value>true</value>

    ?</property>

    ?<property>

    ?? <name>dfs.permissions.superusergroup</name>

    ?? <value>staff</value>

    ?</property>

    ?<property>

    ?? <name>dfs.permissions.enabled</name>

    ?? <value>false</value>

    ?</property>

    </configuration>

    3、yarn-site.xml

    <configuration>

    <!-- Site specific YARN configuration properties -->

    ?<property>

    ?? <name>yarn.resourcemanager.hostname</name>

    ?? <value>centos</value>

    ?</property>

    ?<property>

    ?? <name>yarn.nodemanager.aux-services</name>

    ?? <value>mapreduce_shuffle</value>

    ?</property>

    ?<property>

    ?? <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>

    ?? <value>org.apache.hadoop.mapred.ShuffleHandler</value>

    ?</property>

    ?<property>

    ?? <name>yarn.resourcemanager.address</name>

    ?? <value>centos:18040</value>

    ?</property>

    ?<property>

    ?? <name>yarn.resourcemanager.scheduler.address</name>

    ?? <value>centos:18030</value>

    ?</property>

    ?<property>

    ?? <name>yarn.resourcemanager.resource-tracker.address</name>

    ?? <value>centos:18025</value>

    ?</property>

    ? <property>

    ?? <name>yarn.resourcemanager.admin.address</name>

    ?? <value>centos:18141</value>

    ?</property>

    ?<property>

    ?? <name>yarn.resourcemanager.webapp.address</name>

    ?? <value>centos:18088</value>

    ?</property>

    ?<property>

    ?? <name>yarn.log-aggregation-enable</name>

    ?? <value>true</value>

    ?</property>

    ?<property>

    ?? <name>yarn.log-aggregation.retain-seconds</name>

    ?? <value>86400</value>

    ?</property>

    ?<property>

    ?? <name>yarn.log-aggregation.retain-check-interval-seconds</name>

    ?? <value>86400</value>

    ?</property>

    ?<property>

    ?? <name>yarn.nodemanager.remote-app-log-dir</name>

    ?? <value>/tmp/logs</value>

    ?</property>

    ?<property>

    ?? <name>yarn.nodemanager.remote-app-log-dir-suffix</name>

    ?? <value>logs</value>

    ?</property>

    </configuration>

    4、mapred-site.xml

    # cp mapred-site.xml.template mapred-site.xml

    # vi mapred-site.xml

    <configuration>

    ?<property>

    ? <name>mapreduce.framework.name</name>

    ? <value>yarn</value>

    ?</property>

    ?<property>

    ? <name>mapreduce.jobtracker.http.address</name>

    ? <value>centos:50030</value>

    ?</property>

    ?<property>

    ? <name>mapreduce.jobhisotry.address</name>

    ? <value>centos:10020</value>

    ?</property>

    ?<property>

    ? <name>mapreduce.jobhistory.webapp.address</name>

    ? <value>centos:19888</value>

    ?</property>

    ?<property>

    ? <name>mapreduce.jobhistory.done-dir</name>

    ? <value>/jobhistory/done</value>

    ?</property>

    ?<property>

    ? <name>mapreduce.intermediate-done-dir</name>

    ? <value>/jobhisotry/done_intermediate</value>

    ?</property>

    ?<property>

    ? <name>mapreduce.job.ubertask.enable</name>

    ? <value>true</value>

    ?</property>

    </configuration>

    5、slaves

    centos

    6、hadoop-env.sh

    # The java implementation to use.

    export JAVA_HOME=/usr/lib/jvm/java-1.8.0-openjdk-1.8.0.191.b12-0.el7_5.x86_64

    #修改/etc/profile

    export JAVA_HOME=/usr/lib/jvm/java-1.8.0-openjdk-1.8.0.102-4.b14.el7.x86_64

    export PATH=$PATH:$JAVA_HOME/bin:$HADOOP_HOME/bin:$HADOOP_HOME/sbin

    export CLASSPATH=.:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar

    export HADOOP_HOME=/opt/hadoop/hadoop-2.7.2

    #使生效

    source /etc/profile

    # /opt/hadoop/hadoop-2.7.2/bin/hdfs namenode -format (格式化HDFS)

    # /opt/hadoop/hadoop-2.7.2/sbin/start-all.sh (启动Hadoop集群)

    # jps(验证Hadoop集群)

    2401 ResourceManager

    4066 RunJar

    1798 SecondaryNameNode

    1497 NameNode

    2506 NodeManager

    3244 RunJar

    4783 Jps

    #关闭防火墙,或者在防火墙的规则中开放这些端口

    http://192.168.236.167:50070 (hdfs)

    http://192.168.236.167:18088 (yarn)

    cs