• hadoop beginning


    1、安装虚拟机

    VMware workstation

    CentOS 镜像

    安装Linux虚拟机:(在Win7上)

    1)安装VMwareWorkstations(可修改配置)

    2)添加CentOS镜像(启动安装 ->配置网络)

    网络配置:NAT模式

    网络重启:service network restart

    关闭系统:shutdown -h now

    2、远程连接

    Xshell5

    Xftp5

    3、在Linux上搭建Hadoop集群

    下载软件

    1. Jdk:http://www.oracle.com/technetwork/java/javase/downloads/jdk8-downloads-2133151.html

    2. Hadoop包:http://apache.fayea.com/hadoop/common/hadoop-2.7.2/

    步骤:

    1、修改/etc/hosts

    2、配置JDK

    安装:rpm -ivh jdk-8u101-linux-x64.rpm

    vi /etc/profile
    JAVA_HOME
    source /etc/profile
    java -version

    3、配置SSH(免密码登录)

    ssh-keygen -t rsa
    cat xxx.pub >> authorized_keys
    chmod 644 authorized_keys
    ssh IP/HOSTNAME

    4、安装及配置Hadoop

    tar zxf hadoop-2.7.2.tar.gz
    cd /opt/hadoop-2.7.2/etc/hadoop/

    JAVA_HOME=/usr/java/default
    HADOOP_HOME=/opt/hadoop-2.7.7

    PATH=$JAVA_HOME/bin:$HADOOP_HOME/bin:$PATH

     

    core-site.xml

    <property>
        <name>fs.default.name</name>
        <value>hdfs://bigdata:9000</value>
      </property>
     
     <property>
        <name>hadoop.tmp.dir</name>
        <value>/opt/hadoop-2.7.2/current/tmp</value>
      </property>
     <property>
        <name>fs.trash.interval</name>
        <value>4320</value>
      </property>

    hdfs-site.xml

    <property>
       <name>dfs.namenode.name.dir</name>
       <value>/opt/hadoop-2.7.2/current/dfs/name</value>
     </property>
     <property>
       <name>dfs.datanode.data.dir</name>
       <value>/opt/hadoop-2.7.2/current/data</value>
     </property>
     <property>
       <name>dfs.replication</name>
       <value>1</value>
     </property>
     <property>
       <name>dfs.webhdfs.enabled</name>
       <value>true</value>
     </property>
     <property>
       <name>dfs.permissions.superusergroup</name>
       <value>staff</value>
     </property>
     <property>
       <name>dfs.permissions.enabled</name>
       <value>false</value>
     </property>

    yarn-site.xml

    <property>
       <name>yarn.resourcemanager.hostname</name>
       <value>bigdata</value>
     </property>
     <property>
       <name>yarn.nodemanager.aux-services</name>
       <value>mapreduce_shuffle</value>
     </property>
     <property>
       <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
       <value>org.apache.hadoop.mapred.ShuffleHandler</value>
     </property>
     <property>
       <name>yarn.resourcemanager.address</name>
       <value>bigdata:18040</value>
     </property>
    <property>
       <name>yarn.resourcemanager.scheduler.address</name>
       <value>bigdata:18030</value>
     </property>
     <property>
       <name>yarn.resourcemanager.resource-tracker.address</name>
       <value>bigdata:18025</value>
     </property> <property>
       <name>yarn.resourcemanager.admin.address</name>
       <value>bigdata:18141</value>
     </property>
    <property>
       <name>yarn.resourcemanager.webapp.address</name>
       <value>bigdata:18088</value>
     </property>
    <property>
       <name>yarn.log-aggregation-enable</name>
       <value>true</value>
     </property>
    <property>
       <name>yarn.log-aggregation.retain-seconds</name>
       <value>86400</value>
     </property>
    <property>
       <name>yarn.log-aggregation.retain-check-interval-seconds</name>
       <value>86400</value>
     </property>
    <property>
       <name>yarn.nodemanager.remote-app-log-dir</name>
       <value>/tmp/logs</value>
     </property>
    <property>
       <name>yarn.nodemanager.remote-app-log-dir-suffix</name>
       <value>logs</value>
     </property>

    mapred-site.xml

    <property>
      <name>mapreduce.framework.name</name>
      <value>yarn</value>
    </property>
    <property>
      <name>mapreduce.jobtracker.http.address</name>
      <value>bigdata:50030</value>
    </property>
    <property>
      <name>mapreduce.jobhisotry.address</name>
      <value>bigdata:10020</value>
    </property>
    <property>
      <name>mapreduce.jobhistory.webapp.address</name>
      <value>bigdata:19888</value>
    </property>
    <property>
      <name>mapreduce.jobhistory.done-dir</name>
      <value>/jobhistory/done</value>
    </property>
    <property>
      <name>mapreduce.intermediate-done-dir</name>
      <value>/jobhisotry/done_intermediate</value>
    </property>
    <property>
      <name>mapreduce.job.ubertask.enable</name>
      <value>true</value>
    </property>

    slaves

    bigdata

    hadoop-env.sh

    JAVA_HOME

    5、格式化HDFS

    hdfs namenode -format

    ‘16/09/0403:07:30 INFO common.Storage: Storage directory /opt/hadoop-2.7.2/current/dfs/namehas been successfully formatted.

     

    6、启动Hadoop集群

    /opt/hadoop-2.7.2/sbin/start-all.sh

     

    7、验证Hadoop集群

    1)jps

    2)

    关闭防火墙 或者 在防火墙的规则中开放这些端口

    hdfs   http://bigdata:50070

    yarn  http://bigdata:18088

    HDFS格式化后启动dfs出现以下错误:

    [root@master sbin]# ./start-dfs.sh
    Starting namenodes on [master]
    ERROR: Attempting to operate on hdfs namenode as root
    ERROR: but there is no HDFS_NAMENODE_USER defined. Aborting operation.
    Starting datanodes
    ERROR: Attempting to operate on hdfs datanode as root
    ERROR: but there is no HDFS_DATANODE_USER defined. Aborting operation.
    Starting secondary namenodes [slave1]
    ERROR: Attempting to operate on hdfs secondarynamenode as root
    ERROR: but there is no HDFS_SECONDARYNAMENODE_USER defined. Aborting operation.
    1
    2
    3
    4
    5
    6
    7
    8
    9
    10
    11
    查度娘,见一仁兄的博客有次FAQ,故参考处理顺便再做一记录
    参考地址:https://blog.csdn.net/u013725455/article/details/70147331

    在/hadoop/sbin路径下:
    将start-dfs.sh,stop-dfs.sh两个文件顶部添加以下参数

    #!/usr/bin/env bash
    HDFS_DATANODE_USER=root
    HADOOP_SECURE_DN_USER=hdfs
    HDFS_NAMENODE_USER=root
    HDFS_SECONDARYNAMENODE_USER=root
    1
    2
    3
    4
    5
    6
    还有,start-yarn.sh,stop-yarn.sh顶部也需添加以下:

    #!/usr/bin/env bash
    YARN_RESOURCEMANAGER_USER=root
    HADOOP_SECURE_DN_USER=yarn
    YARN_NODEMANAGER_USER=root


    # Licensed to the Apache Software Foundation (ASF) under one or more
    1
    2
    3
    4
    5
    6
    7
    8
    修改后重启 ./start-dfs.sh,成功!
    ————————————————
    版权声明:本文为CSDN博主「空空dream」的原创文章,遵循CC 4.0 BY-SA版权协议,转载请附上原文出处链接及本声明。
    原文链接:https://blog.csdn.net/lglglgl/article/details/80553828

  • 相关阅读:
    牛客网暑期ACM多校训练营(第一场)
    ACM-ICPC-Swiss-Subregional 2017 训练日志
    2016 全国多校第二场 训练日志
    2018湖南多校第八场 训练日志
    2017 全国多校第九场 训练日志
    2017 全国多校第十场 训练日志
    2014 Kuala Lumpur 训练日志
    2018湖南多校第五场-20180506 训练日志
    2017icpc 花莲训练日志
    面试总结
  • 原文地址:https://www.cnblogs.com/SeaTop/p/11887611.html
Copyright © 2020-2023  润新知