• Hadoop集群搭建


    角色 IP
    master 192.168.10.10
    save-1 192.168.10.11
    save-2 192.168.10.12

    jdk安装,三台机器一样装

    #tar xf jdk-8u161-linux-x64.tar.gz -C /usr/local/

    #mv /usr/local/{jdk1.8.0_161,jdk} #vim /etc/profile.d/jdk.sh export JAVA_HOME=/usr/local/jdk export CLASSPATH=.:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar export PATH=$JAVA_HOME/bin:$PATH #exec bash

      hosts 本地文件解析,三台机器一样

    # vim /etc/hosts
    
    127.0.0.1   localhost localhost.localdomain localhost4 localhost4.localdomain4
    ::1         localhost localhost.localdomain localhost6 localhost6.localdomain6
    192.168.10.10 hadoop-1 masters
    192.168.10.11 hadoop-2 slaves-1
    192.168.10.12 hadoop-3 slaves-2
    

      配置免密码登录

    主节点
    ssh-keygen  一路回车
    ssh-copy-id hadoop-2
    ssh-copy-id hadoop-3
    1从节点配置
    ssh-keygen  一路回车
    ssh-copy-id hadoop-1
    ssh-copy-id hadoop-3
    2从节点配置
    ssh-keygen  一路回车
    ssh-copy-id hadoop-1
    ssh-copy-id hadoop-2
    

      安装hadoop

    tar xf hadoop-1.2.1-bin.tar.gz -C /usr/local/src/
    cd /usr/local/src/hadoop-1.2.1
    mkdir tmp
    cd conf/
    # vim masters 
    masters
    # vim slaves 
    slaves-1
    slaves-2
    [root@masters conf]# vim core-site.xml
    
    <?xml version="1.0"?>
    <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
    
    <!-- Put site-specific property overrides in this file. -->
    
    <configuration>
            <propert>
                    <name>hadoop.tmp.dir</name>
                    <value>/usr/local/src/hadoop-1.2.1/tmp</value>
            </propert>
            <propert>
                    <name>fs.default.name</name>
                    <value>hdfs://192.168.10.10:9000</value>
            </propert>
    </configuration>
    
    
    [root@masters conf]# vim mapred-site.xml 
    
    <?xml version="1.0"?>
    <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
    
    <!-- Put site-specific property overrides in this file. -->
    
    <configuration>
            <propert>
                    <name>mapred.job.tracker</name>
                    <value>http://192.168.10.10:9001</value>
            </propert>
    
    </configuration>
    
    
    [root@masters conf]# vim hdfs-site.xml 
    
    <?xml version="1.0"?>
    <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
    
    <!-- Put site-specific property overrides in this file. -->
    
    <configuration>
            <propert>
                    <name>dfs.replication</name>
                    <value>3</value>
            </propert>
    
    </configuration>
    
    # vim hadoop-env.sh 
    
    export JAVA_HOME=/usr/local/jdk
    

      将hadoop拷贝从节点

    [root@masters conf]# cd ../../
    [root@masters src]# scp -r hadoop-1.2.1 hadoop-2:/usr/local/src/
    [root@masters src]# scp -r hadoop-1.2.1 hadoop-3:/usr/local/src/
    [root@masters src]# cd hadoop-1.2.1/bin/
    [root@masters bin]# ./start-all.sh 
    

      

      

    草都可以从石头缝隙中长出来更可况你呢
  • 相关阅读:
    排序算法<四>希尔排序
    排序算法<三>快速排序
    排序算法<二>冒泡排序
    排序算法<一>选择排序
    相信我这是水题ny1313
    斐波那契查找
    插值查找
    第四届蓝桥杯c/c++B组4
    如何使用NPM来管理你的Node.js依赖
    javascript深入学习
  • 原文地址:https://www.cnblogs.com/rdchenxi/p/11105448.html
Copyright © 2020-2023  润新知