• hadoop-2.0.0-cdh4.6.0、sqoop-1.4.3-cdh4.6.0、mahout-0.7-cdh4.6.0 安装笔记


    1. /etc/profile中环境变量配置:

    export HADOOP_HOME=/usr/local/hadoop/cdh4.6.0/hadoop-2.0.0
    export HADOOP_MAPRED_HOME=${HADOOP_HOME}
    export HADOOP_COMMON_HOME=${HADOOP_HOME}
    export HADOOP_HDFS_HOME=${HADOOP_HOME}
    export YARN_HOME=${HADOOP_HOME}
    export HADOOP_YARN_HOME=${HADOOP_HOME}
    export HADOOP_CONF_DIR=${HADOOP_HOME}/etc/hadoop
    export HDFS_CONF_DIR=${HADOOP_HOME}/etc/hadoop
    export YARN_CONF_DIR=${HADOOP_HOME}/etc/hadoop
    export PATH=$PATH:$HADOOP_HOME/sbin:$HADOOP_HOME/bin
    
    export CLASSPATH=$CLASSPATH:$HADOOP_CONF_DIR
    
    export SQOOP_HOME=/usr/local/hadoop/cdh4.6.0/sqoop-1.4.3
    export PATH=$PATH:$SQOOP_HOME/bin
    
    export MAHOUT_HOME=/usr/local/hadoop/cdh4.6.0/mahout-0.7
    export PATH=$PATH:$MAHOUT_HOME/bin
    环境变量

    1.1  /etc/hosts 中加上:192.168.6.91    ds

    2. $HADOOP_HOME/etc/hadoop目录下,修改配置文件

      1)core-site.xml:

    <configuration>
      <property>
        <name>fs.default.name</name>
        <value>hdfs://ds:9000</value>
      </property>
      <property>
        <name>fs.trash.interval</name>
        <value>10080</value>
      </property>
      <property>
        <name>fs.trash.checkpoint.interval</name>
        <value>10080</value>
      </property>
    </configuration>
    core-site.xml

      2)yarn-site.xml:

    <configuration>
    <!-- Site specific YARN configuration properties -->
      <property>
        <name>yarn.resourcemanager.resource-tracker.address</name>
        <value>ds:8031</value>
      </property>
      <property>
        <name>yarn.resourcemanager.address</name>
        <value>ds:8032</value>
      </property>
      <property>
        <name>yarn.resourcemanager.scheduler.address</name>
        <value>ds:8030</value>
      </property>
      <property>
        <name>yarn.resourcemanager.admin.address</name>
        <value>ds:8033</value>
      </property>
      <property>
        <name>yarn.resourcemanager.webapp.address</name>
        <value>ds:8088</value>
      </property>
      <property>
        <description>Classpath for typical applications.</description>
        <name>yarn.application.classpath</name>
        <value>$HADOOP_CONF_DIR,$HADOOP_COMMON_HOME/share/hadoop/common/*,
        $HADOOP_COMMON_HOME/share/hadoop/common/lib/*,
        $HADOOP_HDFS_HOME/share/hadoop/hdfs/*,$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*,
        $YARN_HOME/share/hadoop/yarn/*,$YARN_HOME/share/hadoop/yarn/lib/*,
        $YARN_HOME/share/hadoop/mapreduce/*,$YARN_HOME/share/hadoop/mapreduce/lib/*
        </value>
      </property>
      <property>
        <name>yarn.nodemanager.aux-services</name>
        <value>mapreduce.shuffle</value>
      </property>
      <property>
        <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
        <value>org.apache.hadoop.mapred.ShuffleHandler</value>
      </property>
      <property>
        <name>yarn.nodemanager.local-dirs</name>
        <value>/usr/local/hadoop/cdh4.6.0/hadoop-2.0.0/yarn/local</value>
      </property>
      <property>
        <name>yarn.nodemanager.log-dirs</name>
        <value>/usr/local/hadoop/cdh4.6.0/hadoop-2.0.0/yarn/logs</value>
      </property>
      <property>
        <description>Where to aggregate logs</description>
        <name>yarn.nodemanager.remote-app-log-dir</name>
        <value>/usr/local/hadoop/cdh4.6.0/hadoop-2.0.0/yarn/logs</value>
      </property>
      <property>
        <name>yarn.app.mapreduce.am.staging-dir</name>
        <value>/usr/local/hadoop/cdh4.6.0/hadoop-2.0.0</value>
    </property>
    </configuration>
    yarn-site.xml

      3)mapred-site.xml:

    <configuration>
     <property>
        <name>mapreduce.framework.name</name>
        <value>yarn</value>
      </property>
      <property>
        <name>mapred.job.tracker</name>
        <value>ds:9101</value>
      </property>
      <property>
        <name>mapreduce.jobhistory.address</name>
        <value>ds:10020</value>
      </property>
      <property>
        <name>mapreduce.jobhistory.webapp.address</name>
        <value>ds:19888</value>
      </property>
    
    
    </configuration>
    mapred-site.xml

    4)hadoop-env.sh

      最后加上export JAVA_HOME=/usr/local/jdk1.6.0_13  或者  export JAVA_HOME=$JAVA_HOME

    3. 创建目录:$HADOOP_HOME/tmp/dfs/data

    4. 启动hadoop:

      1)hadoop namenode -format

      2) start-dfs.sh

      jps看DataNode、SecondaryNameNode、NameNode 是否都起来了,没起来的话查对应的log

      3)start-yarn.sh

      jps看NodeManager、ResourceManager 是否都起来了

    5. 修改$SQOOP_HOME/bin/configure-sqoop

      注释掉所有关于zookeeper和hbase的部分

    6. 命令:sqoop help  如果出现usage: ... 则安装成功

    7. 命令:mahout  如果出现Valid program names are: ... 则Mahout安装成功 (Mahout不需要更改配置文件)

  • 相关阅读:
    菜鸟看懂算法以后之一:头痛的64次左移
    C语言通过指针数组和二维数组读取文件
    C++中构造函数调用构造函数
    bnuoj53075 外挂使用拒绝
    [CodeForces]String Reconstruction
    BNU-2017.7.4排位赛2总结
    BNU-2017.7.5排位赛3总结
    BNU-2017.7.3排位赛1总结
    微软大楼设计方案(困难)
    最长公共子序列针对小字符集的算法
  • 原文地址:https://www.cnblogs.com/-wangjiannan/p/3723049.html
Copyright © 2020-2023  润新知