• linux编译64bitHadoop (eg: ubuntu14.04 and hadoop 2.3.0)


    Hadoop官网提供的编译好的hadoop-2.3.0.tar.gz二进制包是在32位系统上编译的,在64系统上运行会有一些错误,比如:

    WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable

    此时需要自行编译hadoop 2.30 源码。本人编译的hadoop 2.30的二进制包经实验证明可以正确安装并正确运行了Word Count程序。此处可以下载我编译的二进制包:

    http://pan.baidu.com/s/1eQrgsWa

    以下为编译步骤

    1. 更新软件包列表

    apt-get update

    2. 安装编译所需要的软件: 为什么需要这些依赖包? 我也不知道==

    apt-get install -y openjdk-7-jdk libprotobuf-dev protobuf-compiler maven cmake build-essential pkg-config libssl-dev zlib1g-dev llvm-gcc automake autoconf make

    3. 下载hadoop 2.30的源文件包

    wget http://archive.apache.org/dist/hadoop/core/hadoop-2.3.0/hadoop-2.3.0-src.tar.gz

    4. 解压hadoop 2.30 的源文件包

    tar -xzvf hadoop-2.3.0-src.tar.gz

    5. 进入hadoop 2.30 文件夹

    cd hadoop-2.3.0-src

    6. 编译hadoop 2.30 源文件

    mvn package -Pdist,native -DskipTests –Dtar

    正确执行的结果如下:

    [INFO] ------------------------------------------------------------------------

    [INFO] Reactor Summary:

    [INFO]

    [INFO] Apache Hadoop Main ................................ SUCCESS [1:11.968s]

    [INFO] Apache Hadoop Project POM ......................... SUCCESS [30.393s]

    [INFO] Apache Hadoop Annotations ......................... SUCCESS [18.398s]

    [INFO] Apache Hadoop Assemblies .......................... SUCCESS [0.246s]

    [INFO] Apache Hadoop Project Dist POM .................... SUCCESS [20.372s]

    [INFO] Apache Hadoop Maven Plugins ....................... SUCCESS [23.721s]

    [INFO] Apache Hadoop MiniKDC ............................. SUCCESS [1:41.836s]

    [INFO] Apache Hadoop Auth ................................ SUCCESS [22.303s]

    [INFO] Apache Hadoop Auth Examples ....................... SUCCESS [7.052s]

    [INFO] Apache Hadoop Common .............................. SUCCESS [2:29.466s]

    [INFO] Apache Hadoop NFS ................................. SUCCESS [11.604s]

    [INFO] Apache Hadoop Common Project ...................... SUCCESS [0.073s]

    [INFO] Apache Hadoop HDFS ................................ SUCCESS [1:30.230s]

    [INFO] Apache Hadoop HttpFS .............................. SUCCESS [17.976s]

    [INFO] Apache Hadoop HDFS BookKeeper Journal ............. SUCCESS [19.927s]

    [INFO] Apache Hadoop HDFS-NFS ............................ SUCCESS [3.304s]

    [INFO] Apache Hadoop HDFS Project ........................ SUCCESS [0.032s]

    [INFO] hadoop-yarn ....................................... SUCCESS [0.033s]

    [INFO] hadoop-yarn-api ................................... SUCCESS [36.284s]

    [INFO] hadoop-yarn-common ................................ SUCCESS [33.912s]

    [INFO] hadoop-yarn-server ................................ SUCCESS [0.213s]

    [INFO] hadoop-yarn-server-common ......................... SUCCESS [8.193s]

    [INFO] hadoop-yarn-server-nodemanager .................... SUCCESS [41.181s]

    [INFO] hadoop-yarn-server-web-proxy ...................... SUCCESS [2.768s]

    [INFO] hadoop-yarn-server-resourcemanager ................ SUCCESS [13.923s]

    [INFO] hadoop-yarn-server-tests .......................... SUCCESS [0.904s]

    [INFO] hadoop-yarn-client ................................ SUCCESS [4.363s]

    [INFO] hadoop-yarn-applications .......................... SUCCESS [0.120s]

    [INFO] hadoop-yarn-applications-distributedshell ......... SUCCESS [2.262s]

    [INFO] hadoop-yarn-applications-unmanaged-am-launcher .... SUCCESS [1.615s]

    [INFO] hadoop-yarn-site .................................. SUCCESS [0.086s]

    [INFO] hadoop-yarn-project ............................... SUCCESS [2.703s]

    [INFO] hadoop-mapreduce-client ........................... SUCCESS [0.132s]

    [INFO] hadoop-mapreduce-client-core ...................... SUCCESS [18.951s]

    [INFO] hadoop-mapreduce-client-common .................... SUCCESS [14.320s]

    [INFO] hadoop-mapreduce-client-shuffle ................... SUCCESS [3.330s]

    [INFO] hadoop-mapreduce-client-app ....................... SUCCESS [9.664s]

    [INFO] hadoop-mapreduce-client-hs ........................ SUCCESS [7.678s]

    [INFO] hadoop-mapreduce-client-jobclient ................. SUCCESS [9.263s]

    [INFO] hadoop-mapreduce-client-hs-plugins ................ SUCCESS [1.549s]

    [INFO] Apache Hadoop MapReduce Examples .................. SUCCESS [5.748s]

    [INFO] hadoop-mapreduce .................................. SUCCESS [2.880s]

    [INFO] Apache Hadoop MapReduce Streaming ................. SUCCESS [7.080s]

    [INFO] Apache Hadoop Distributed Copy .................... SUCCESS [14.648s]

    [INFO] Apache Hadoop Archives ............................ SUCCESS [2.602s]

    [INFO] Apache Hadoop Rumen ............................... SUCCESS [5.706s]

    [INFO] Apache Hadoop Gridmix ............................. SUCCESS [3.649s]

    [INFO] Apache Hadoop Data Join ........................... SUCCESS [2.483s]

    [INFO] Apache Hadoop Extras .............................. SUCCESS [2.678s]

    [INFO] Apache Hadoop Pipes ............................... SUCCESS [6.359s]

    [INFO] Apache Hadoop OpenStack support ................... SUCCESS [5.088s]

    [INFO] Apache Hadoop Client .............................. SUCCESS [4.534s]

    [INFO] Apache Hadoop Mini-Cluster ........................ SUCCESS [0.433s]

    [INFO] Apache Hadoop Scheduler Load Simulator ............ SUCCESS [7.757s]

    [INFO] Apache Hadoop Tools Dist .......................... SUCCESS [4.099s]

    [INFO] Apache Hadoop Tools ............................... SUCCESS [0.428s]

    [INFO] Apache Hadoop Distribution ........................ SUCCESS [18.045s]

    [INFO] ------------------------------------------------------------------------

    [INFO] BUILD SUCCESS

    [INFO] ------------------------------------------------------------------------

    [INFO] Total time: 14:59.240s

    [INFO] Finished at: Thu Jan 15 18:51:59 JST 2015

    [INFO] Final Memory: 168M/435M

    [INFO] ------------------------------------------------------------------------

    编译好的二进制文件包位于

    hadoop-2.3.0-src/hadoop-dist/target/hadoop-2.3.0.tar.gz

    PS: 使用自行编译的hadoop 2.30二进制包安装hadoop 2.30时需要注意删除 .bashrc文件与hadoop-env.sh文件中下面两行(默认不会有这两行,但是尝试解决报错时可能改写了)

    export HADOOP_COMMON_LIB_NATIVE_DIR="~/hadoop/lib/"

    export HADOOP_OPTS="$HADOOP_OPTS -Djava.library.path=~/hadoop/lib/"

  • 相关阅读:
    loj#2049. 「HNOI2016」网络(set 树剖 暴力)
    创建多个Oracle数据库及相应的实例
    [置顶] lua 进阶3--lua文件中调用C++函数
    android 设置Button或者ImageButton的背景透明 半透明 透明
    struts2 18拦截器详解(七)
    《Linux命令行与shell脚本编程大全》 第二十三章 学习笔记
    ios7下不能录音问题解决
    360 2013校园招聘笔试题(含参考答案)
    【MFC三天一个游戏】之 局域网黑白棋
    Remove Nth Node From End of List
  • 原文地址:https://www.cnblogs.com/felixzh/p/5086736.html
Copyright © 2020-2023  润新知