• Spark记录-源码编译spark2.2.0(结合Hive on Spark/Hive on MR2/Spark on Yarn)


    #spark2.2.0源码编译

    #组件:mvn-3.3.9 jdk-1.8

    #wget http://mirror.bit.edu.cn/apache/spark/spark-2.2.0/spark-2.2.0.tgz ---下载源码 (如果是Hive on spark---hive2.1.1对应spark1.6.0)

    #tar zxvf spark-2.2.0.tgz ---解压

    #cd spark-2.2.0/dev

    ##修改make-distribution.sh的MVN路径为$M2_HOME/bin/mvn   ---查看并安装pom.xml的mvn版本

    ##cd ..   ---切换到spark-2.2.0

    #./dev/change-scala-version.sh 2.11 ---更改scala版本(低于11不需要此步骤)

    #./dev/make-distribution.sh --name "hadoop2-without-hive" --tgz "-Pyarn,hadoop-provided,hadoop-2.7,parquet-provided"    ---生成在根目录下

    ##下载的可直接运行的是2.10.x,源码支持2.11.x,但是编译的时候需要加上条件-Dscala-2.11,但是Spark does not yet support its JDBC component for Scala 2.11. 所以需要这部分功能的还是用2.10.x。

    ##export MAVEN_OPTS="-Xmx2g -XX:MaxPermSize=512M -XX:ReservedCodeCacheSize=512m"

    ##./build/mvn -Pyarn -Phadoop-2.7 -Dhadoop.version=2.7.3 -DskipTests clean package   ---生成在dist目录下

    ##参考http://spark.apache.org/docs/latest/building-spark.html

    ./dev/make-distribution.sh --name custom-spark --pip --r --tgz -Psparkr -Phadoop-2.7 -Phive -Phive-thriftserver -Pmesos -Pyarn

    ##With Hive 1.2.1 support

    ./build/mvn -Pyarn -Phive -Phive-thriftserver -DskipTests clean package

    ##With scala 2.10 support

    ./dev/change-scala-version.sh 2.10

    ./build/mvn -Pyarn -Dscala-2.10 -DskipTests clean package

    ##with Mesos support

    ./build/mvn -Pmesos -DskipTests clean package

    ##build with sbt

    ./build/sbt package

    ##Apache Hadoop 2.6.X

    ./build/mvn -Pyarn -DskipTests clean package

    ##Apache Hadoop 2.7.X and later

    ./build/mvn -Pyarn -Phadoop-2.7 -Dhadoop.version=2.7.3 -DskipTests clean package

    ##参考https://cwiki.apache.org//confluence/display/Hive/Hive+on+Spark:+Getting+Started

    Follow instructions to install Spark:  

     YARN Mode: http://spark.apache.org/docs/latest/running-on-yarn.html 

    #spark-submit

    $ ./bin/spark-submit --class path.to.your.Class --master yarn --deploy-mode cluster [options] <app jar> [app options]

    For example:

    $ ./bin/spark-submit --class org.apache.spark.examples.SparkPi  --master yarn  --deploy-mode cluster  --driver-memory 4g  --executor-memory 2g  --executor-cores 1  --queue thequeue lib/spark-examples*.jar  10

    #spark-shell

    $ ./bin/spark-shell --master yarn --deploy-mode client

    #spark-submit add jars

    $ ./bin/spark-submit --class my.main.Class  --master yarn  --deploy-mode cluster  --jars my-other-jar.jar,my-other-other-jar.jar  my-main-jar.jar  app_arg1 app_arg2

    Standalone Mode: https://spark.apache.org/docs/latest/spark-standalone.html

    #./bin/spark-shell --master spark://IP:PORT   ---您还可以传递一个选项--total-executor-cores <numCores>来控制spark-shell在群集上使用的核心数。

    #./bin/spark-class org.apache.spark.deploy.Client kill <master url> <driver ID>

    Hive on Spark supports Spark on YARN mode as default.

    For the installation perform the following tasks:

      1. Install Spark (either download pre-built Spark, or build assembly from source).  
        • Install/build a compatible version.  Hive root pom.xml's <spark.version> defines what version of Spark it was built/tested with. 
        • Install/build a compatible distribution.  Each version of Spark has several distributions, corresponding with different versions of Hadoop.
        • Once Spark is installed, find and keep note of the <spark-assembly-*.jar> location.
        • Note that you must have a version of Spark which does not include the Hive jars. Meaning one which was not built with the Hive profile. If you will use Parquet tables, it's recommended to also enable the "parquet-provided" profile. Otherwise there could be conflicts in Parquet dependency. To remove Hive jars from the installation, simply use the following command under your Spark repository:

    Prior to Spark 2.0.0:

    ./make-distribution.sh --name "hadoop2-without-hive" --tgz "-Pyarn,hadoop-provided,hadoop-2.4,parquet-provided"

    Since Spark 2.0.0:

    ./dev/make-distribution.sh --name "hadoop2-without-hive" --tgz "-Pyarn,hadoop-provided,hadoop-2.7,parquet-provided"
    1. To add the Spark dependency to Hive:

      • Prior to Hive 2.2.0, link the spark-assembly jar to HIVE_HOME/lib.
      • Since Hive 2.2.0, Hive on Spark runs with Spark 2.0.0 and above, which doesn't have an assembly jar.
        • To run with YARN mode (either yarn-client or yarn-cluster), link the following jars to HIVE_HOME/lib.
          • scala-library
          • spark-core
          • spark-network-common
        • To run with LOCAL mode (for debugging only), link the following jars in addition to those above to HIVE_HOME/lib.
          • chill-java  chill  jackson-module-paranamer  jackson-module-scala  jersey-container-servlet-core
          • jersey-server  json4s-ast  kryo-shaded  minlog  scala-xml  spark-launcher
          • spark-network-shuffle  spark-unsafe  xbean-asm5-shaded
    2. Configure Hive execution engine to use Spark

      set hive.execution.engine=spark;

     

    1. Configure Spark-application configs for Hive.  See: http://spark.apache.org/docs/latest/configuration.html.  This can be done either by adding a file "spark-defaults.conf" with these properties to the Hive classpath, or by setting them on Hive configuration (hive-site.xml). For instance:

      set spark.master=<Spark Master URL>
      set spark.eventLog.enabled=true;
      set spark.eventLog.dir=<Spark event log folder (must exist)>
      set spark.executor.memory=512m;             
      set spark.serializer=org.apache.spark.serializer.KryoSerializer;

      Configuration property details

      • spark.executor.memory: Amount of memory to use per executor process.
      • spark.executor.cores: Number of cores per executor.
      • spark.yarn.executor.memoryOverhead: The amount of off heap memory (in megabytes) to be allocated per executor, when running Spark on Yarn. This is memory that accounts for things like VM overheads, interned strings, other native overheads, etc. In addition to the executor's memory, the container in which the executor is launched needs some extra memory for system processes, and this is what this overhead is for.

      • spark.executor.instances: The number of executors assigned to each application.
      • spark.driver.memory: The amount of memory assigned to the Remote Spark Context (RSC). We recommend 4GB.
      • spark.yarn.driver.memoryOverhead: We recommend 400 (MB).
    2. Allow Yarn to cache necessary spark dependency jars on nodes so that it does not need to be distributed each time when an application runs.

      • Prior to Hive 2.2.0, upload spark-assembly jar to hdfs file(for example: hdfs://xxxx:8020/spark-assembly.jar) and add following in hive-site.xml

        <property>
          <name>spark.yarn.jar</name>
          <value>hdfs://xxxx:8020/spark-assembly.jar</value>
        </property>
      • Hive 2.2.0, upload all jars in $SPARK_HOME/jars to hdfs folder(for example:hdfs:///xxxx:8020/spark-jars) and add following in hive-site.xml

        <property>
          <name>spark.yarn.jars</name>
          <value>hdfs://xxxx:8020/spark-jars/*</value>
        </property>

    Configuring Spark

    Setting executor memory size is more complicated than simply setting it to be as large as possible. There are several things that need to be taken into consideration:

    • More executor memory means it can enable mapjoin optimization for more queries.

    • More executor memory, on the other hand, becomes unwieldy from GC perspective.

    • Some experiments shows that HDFS client doesn’t handle concurrent writers well, so it may face race condition if executor cores are too many. 

    The following settings need to be tuned for the cluster, these may also apply to submission of Spark jobs outside of Hive on Spark:

    Property
    Recommendation
    spark.executor.cores Between 5-7, See tuning details section
    spark.executor.memory yarn.nodemanager.resource.memory-mb * (spark.executor.cores / yarn.nodemanager.resource.cpu-vcores) 
    spark.yarn.executor.memoryOverhead 15-20% of spark.executor.memory
    spark.executor.instances Depends on spark.executor.memory + spark.yarn.executor.memoryOverhead, see tuning details section.

     #参考http://spark.apache.org/docs/latest/configuration.html

    ##maven阿里仓库settings.xml

    <mirror>
        <id>nexus-aliyun</id>
        <mirrorOf>*</mirrorOf>
        <name>Nexus aliyun</name>
        <url>http://maven.aliyun.com/nexus/content/groups/public</url>
    </mirror>
    pom.xml
    <repositories>
            <repository>
                <id>nexus-aliyun</id>
                <name>Nexus aliyun</name>
                <url>http://maven.aliyun.com/nexus/content/groups/public</url>
            </repository>
        </repositories>
    <!--<url>https://repo1.maven.org/maven2</url>-->    <url>http://maven.aliyun.com/nexus/content/groups/public/</url>

    #大约1个多小时,保证联网环境,下载和编译包需要时间,如遇到问题请自查,重新执行上述命令

    #官网下载scala2.10.5解压,并命令为scala

    #vim /etc/profile

    export JAVA_HOME=/usr/app/jdk1.8.0 
    export SCALA_HOME=/usr/app/scala 
    export HADOOP_HOME=/usr/app/hadoop
    export HADOOP_CONF_DIR=${HADOOP_HOME}/etc/hadoop  
    export HADOOP_COMMON_LIB_NATIVE_DIR=${HADOOP_HOME}/lib/native  
    export HADOOP_OPTS="-Djava.library.path=${HADOOP_HOME}/lib"  
    export HIVE_HOME=/usr/app/hive 
    export HIVE_CONF_DIR=${HIVE_HOME}/conf 
    export HBASE_HOME=/usr/app/hbase
    export HBASE_CONF_DIR=${HBASE_HOME}/conf  
    export SPARK_HOME=/usr/app/spark 
    export CLASSPATH=.:${JAVA_HOME}/lib:${SCALA_HOME}/lib:${HIVE_HOME}/lib:${HBASE_HOME}/lib:$CLASSPATH  
    export PATH=.:${JAVA_HOME}/bin:${SCALA_HOME}/bin:${HADOOP_HOME}/bin:${HADOOP_HOME}/sbin:${SPARK_HOME}/bin:${HIVE_HOME}/bin:${HBASE_HOME}/bin:$PATH 

    #spark-env.sh

    export JAVA_HOME=/usr/app/jdk1.8.0
    export SCALA_HOME=/usr/app/scala
    export HADOOP_HOME=/usr/app/hadoop
    export HADOOP_CONF_DIR=/usr/app/hadoop/etc/hadoop
    export YARN_CONF_DIR=/usr/app/hadoop/etc/hadoop
    #export SPARK_LAUNCH_WITH_SCALA=0
    #export SPARK_WORKER_MEMORY=512m
    #export SPARK_DRIVER_MEMORY=512m
    export SPARK_EXECUTOR_MEMORY=512M  
    export SPARK_MASTER_IP=192.168.66.66
    export SPARK_HOME=/usr/app/spark
    export SPARK_LIBRARY_PATH=/usr/app/spark/lib
    export SPARK_MASTER_WEBUI_PORT=18080
    export SPARK_WORKER_DIR=/usr/app/spark/work
    export SPARK_MASTER_PORT=7077
    export SPARK_WORKER_PORT=7078
    export SPARK_LOG_DIR=/usr/app/spark/logs
    export SPARK_PID_DIR='/usr/app/spark/run'
    export SPARK_DIST_CLASSPATH=$(/usr/app/hadoop/bin/hadoop classpath) 

    #spark-default.conf

    spark.master                     spark://192.168.66.66:7077
    #spark.home                       /usr/app/spark
    spark.eventLog.enabled           true
    spark.eventLog.dir               hdfs://xinfang:9000/spark-log
    spark.serializer                 org.apache.spark.serializer.KryoSerializer
    #spark.executor.memory            512m
    spark.driver.memory              700m
    spark.executor.extraJavaOptions  -XX:+PrintGCDetails -Dkey=value -Dnumbers="one two three"
    #spark.driver.extraJavaOptions    -XX:PermSize=128M -XX:MaxPermSize=256M
    spark.yarn.jar                   hdfs://192.168.66.66:9000/spark-assembly-1.6.0-hadoop2.4.0.jar 

    spark.master指定Spark运行模式,可以是yarn-client、yarn-cluster、spark://xinfang:7077

    spark.home指定SPARK_HOME路径

    spark.eventLog.enabled需要设为true

    spark.eventLog.dir指定路径,放在master节点的hdfs中,端口要跟hdfs设置的端口一致(默认为8020),否则会报错

    spark.executor.memory和spark.driver.memory指定executor和dirver的内存,512m或1g,既不能太大也不能太小,因为太小运行不了,太大又会影响其他服务

    #配置yarn-site.xml

    <configuration>
    <property>
    <name>yarn.nodemanager.aux-services</name>
    <value>mapreduce_shuffle</value>
    </property>
    <property>
      <name>yarn.resourcemanager.scheduler.class</name>
     <value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.FairScheduler</value>
    </property>
    <property>  
           <name>yarn.scheduler.maximum-allocation-mb</name>  
           <value>8182</value>      
    </property>  
    <property>  
           <name>yarn.nodemanager.vmem-pmem-ratio</name>  
           <value>3.1</value>  
    </property>  
    <property>  
           <name>yarn.nodemanager.resource.memory-mb</name>  
           <value>2048</value>  
    </property>  
    <property>
        <name>yarn.nodemanager.pmem-check-enabled</name>
        <value>false</value>
    </property>
    <property>
        <name>yarn.nodemanager.vmem-check-enabled</name>
        <value>false</value>
    </property>
    </configuration>

    说明:yarn.nodemanager.vmem-check-enabled 这个配置的意思是忽略虚拟内存的检查,如果你是安装在虚拟机上,这个配置很有用,配上去之后后续操作不容易出问题。如果是实体机上,并且内存够多,可以将这个配置去掉。

    #低版本:将spark/lib/spark-assembly-*.jar拷贝到$HIVE_HOME/lib目录下)

    #高版本:将spark/jars/spark*.jar和scala*jar拷贝到$HIVE_HOME/lib目录下)

    #hadoop dfs -put /usr/app/spark/lib/spark-assembly-1.6.0-hadoop2.4.0.jar /    ---上传jar文件 ---这一点非常关键

    #配置hive-site.xml

    <configuration>
    <property>  
    <name>hive.metastore.schema.verification</name>  
    <value>false</value>  
    </property>
    <property>
    <name>javax.jdo.option.ConnectionURL</name>
    <value>jdbc:mysql://192.168.66.66:3306/hive?createDatabaseIfNotExist=true</value>
    </property>
    <property>
    <name>javax.jdo.option.ConnectionDriverName</name>
    <value>com.mysql.jdbc.Driver</value>
    </property>
    <property>
    <name>javax.jdo.option.ConnectionUserName</name>
    <value>hive</value>
    </property>
    <property>
    <name>javax.jdo.option.ConnectionPassword</name>
    <value>1</value>
    </property>
    <!--<property>
    <name>hive.hwi.listen.host</name>
    <value>192.168.66.66</value>
    </property>
    <property>
    <name>hive.hwi.listen.port</name>
    <value>9999</value>
    </property>
    <property>
    <name>hive.hwi.war.file</name>
    <value>lib/hive-hwi-2.1.1.war</value>
    </property>-->
    <property>
    <name>hive.metastore.warehouse.dir</name>
    <value>/usr/hive/warehouse</value>
    </property>
    <property>
    <name>hive.exec.scratchdir</name>
    <value>/usr/hive/tmp</value>
    </property>
    <property>
    <name>hive.querylog.location</name>
    <value>/usr/hive/log</value>
    </property>
    <property>  
    <name>hive.server2.thrift.port</name>  
    <value>10000</value>
    </property> 
    <property>  
    <name>hive.server2.thrift.bind.host</name>  
    <value>192.168.66.66</value>
    </property>
    <property>
    <name>hive.server2.webui.host</name>
    <value>192.168.66.66</value>
    </property>
    <property>
    <name>hive.server2.webui.port</name>
    <value>10002</value>
    </property> 
    <property>  
    <name>hive.server2.long.polling.timeout</name>  
    <value>5000</value>                                
    </property>
    <property> 
    <name>hive.server2.enable.doAs</name> 
    <value>true</value> 
    </property> 
    <property> 
    <name>datanucleus.autoCreateSchema</name> 
    <value>false</value> 
    </property> 
    <property> 
    <name>datanucleus.fixedDatastore</name> 
    <value>true</value> 
    </property>
    <!-- hive on mr-->
    <!--
    <property>  
    <name>mapred.job.tracker</name>  
    <value>http://192.168.66.66:9001</value>  
    </property>
    <property>  
    <name>mapreduce.framework.name</name>  
    <value>yarn</value>  
    </property>
    -->
    <!--hive on spark or spark on yarn -->
    <property>  
    <name>hive.enable.spark.execution.engine</name>  
    <value>true</value>  
    </property>  
    <property>
    <name>hive.execution.engine</name>
    <value>spark</value>
    </property>
    <property>
    <name>spark.home</name>
    <value>/usr/app/spark</value>
    </property>
    <property>
    <name>spark.master</name>
    <value>yarn-client</value>
    </property>
    <property>  
    <name>spark.submit.deployMode</name>  
    <value>client</value>  
    </property> 
    <property>
    <name>spark.eventLog.enabled</name>
    <value>true</value>
    </property>
    <!--
    <property>  
    <name>spark.yarn.jar</name>  
    <value>hdfs://192.168.66.66:9000/spark-assembly-1.6.0-hadoop2.4.0.jar</value>  
    </property> 
    <property>  
    <name>spark.executor.cores</name>  
    <value>1</value>  
    </property>--> 
    <property>
    <name>spark.eventLog.dir</name>
    <value>hdfs://192.168.66.66:9000/spark-log</value>
    </property>
    <property>
    <name>spark.serializer</name>
    <value>org.apache.spark.serializer.KryoSerializer</value>
    </property>
    <property>
    <name>spark.executor.memeory</name>
    <value>512m</value>
    </property>
    <!--
    <property>
    <name>spark.driver.memeory</name>
    <value>512m</value>
    </property>-->
    <property>
    <name>spark.executor.extraJavaOptions</name>
    <value>-XX:+PrintGCDetails -Dkey=value -Dnumbers="one two three"</value>
    </property>
    </configuration>

     #拷贝hive-site.xml到spark/conf下(这点非常关键)

    #新建目录

    hadoop fs  -mkdir  -p   /spark-log
    hadoop   fs  -chmod  777  /spark-log
    mkdir -p  /usr/app/spark/work  /usr/app/spark/logs  /usr/app/spark/run
    mkdir -p /usr/app/hive/logs
    
    chmod -R 755 /usr/app/hive  /usr/app/scala /usr/app/spark
    

    #hive 其他参考配置

    1.配置 hive2.2.0(前提要配置hadoop2.7.2,前面文档有介绍)
    #官网下载二进制包,解压到/usr/app 下,配置/etc/profile:
    export HIVE_HOME=/usr/app/hive
    export PATH=$PATH:$HIVE_HOME/bin
    #配置 hive/conf
    #hive-env.sh加入
    #export HADOOP_HEAPSIZE=1024
    export HADOOP_HOME=/usr/app/hadoop
    export HIVE_CONF_DIR=/usr/app/hive/conf
    export HIVE_AUX_JARS_PATH=/usr/app/hive/lib
    #source /etc/profile 立即生效
    #新建目录
    hdfs dfs -mkdir -p /usr/hive/warehouse
    hdfs dfs -mkdir -p /usr/hive/tmp
    hdfs dfs -mkdir -p /usr/hive/log
    hdfs dfs -chmod o+rwx /usr/hive/warehouse
    hdfs dfs -chmod o+rwx /usr/hive/tmp
    hdfs dfs -chmod o+rwx /usr/hive/log
    #配置日志目录
    mkdir -p /usr/app/hive/logs
    conf/hive-log4j.properties修改:
    hive.log.dir=/usr/app/hive/logs
    2.配置Mysql
    #安装mysql->yum -y install mysql-devel mysql-server
    #根据实际调整/etc/my.cnf配置,找不到可通过locate my.cnf查找 
    #cp /usr/share/mysql/my-medium.cnf  /etc/my.cnf
    #启动service mysqld start   /restart/stop
    #进行mysql授权操作
    #mysql>grant all privileges on *.* to root@'%' identified by '1' with grant option;
    #mysql>gant all privileges on *.* to root@'localhost' identified by '1' with grant option;
    #mysql>flush privileges;
    #mysql>exit
    3.mysql 新建 hive 数据库
    #mysql -uroot -p
    #输入密码
    #mysql>create database hive;
    #mysql>alter database hive character set latin1;
    #mysql>grant all privileges on hive.* to hive@'%' identified by '1';
    #mysql>gant all privileges on *.* to hive@'localhost' identified by '1';
    #mysql>flush privileges;
    #exit
    4.编译hive war(web接口)#下载hive2.1.1 src源码解压切换到 hive/hwi/web 执行 jar cvf hive-hwi-2.1.1.war ./* 拷贝到hive/lib 下
    5.修改配置
    修改hadoop -core-site.xml加入
    <property>
    <name>hadoop.proxyuser.root.hosts</name>
    <value>*</value>
    </property>
    <property>
    <name>hadoop.proxyuser.root.groups</name>
    <value>*</value>
    </property>
    6.初始化数据库(本文用的是mysql)
    #$HIVE_HOME/bin/schematool -initSchema -dbType mysql
    7.根据实际情况是否拷贝 jar 到 hive/lib 下(包可到网上搜索下载)
    #拷贝 hadoop/lib/hadoop-lzo-0.4.21-SNAPSHOT.jar 到 hive/lib
    #拷贝 mysql-connector-java-5.1.34.jar 到 hive/lib
    #拷贝 jasper-compiler-5.5.23.jar jasper-runtime-5.5.23.jar commons-el-5.5.23.jar 到 hive/lib
    #拷贝 ant/lib/ant-1.9.4.jar ant-launcher.jar 到 hive/lib(如果系统安装有ant就需要调整ant)
    #如果启动包日志包重复需要删除
    #根据实际修改hive/bin/hive:(根据spark2后的包分散了)
    sparkAssemblyPath='ls ${SPARK_HOME}/lib/spark-assembly-*.jar' 
    将其修改为:sparkAssemblyPath='ls ${SPARK_HOME}/jars/*.jar'
    8.启动 hive
    #先启动hadoop
    #hive --service metastore
    #hive --service hiveserver2  #http://192.168.66.66:10002  进入hiveserver2服务
    #netstat -nl |grep 10000 
    #hive #进入终端
    #hive --service hwi #进入 hive web 页面http://192.168.66.66:9999/hwi/
    

    #Hive on Spark/Spark on Yarn测试(以Hive2.1.1与spark1.6.0为例,需要重新编译spark1.6.0---不能包括Hive)

    $ssh 192.168.66.66

    $sh /usr/app/hadoop/sbin/start-all.sh    ---启动hadoop

    $sh /usr/app/spark/sbin/start-all.sh       ---启动spark

    #Hive on Spark=>Spark on Yarn模式

    $hive     ---进入Hive客户端

    hive>set hive.execution.engine=spark;

    hive>create database test;

    hive>create table test.test(id int,name string);

    hive>insert into test.test(id,name) values('1','china');

    ###hive>select count(*) from test.test;

    Hive客户端

    hive> select count(*) from test.test;
    Query ID = root_20171106100849_3068b783-b729-4d7a-9353-ffc7e29b3685
    Total jobs = 1
    Launching Job 1 out of 1
    In order to change the average load for a reducer (in bytes):
      set hive.exec.reducers.bytes.per.reducer=<number>
    In order to limit the maximum number of reducers:
      set hive.exec.reducers.max=<number>
    In order to set a constant number of reducers:
      set mapreduce.job.reduces=<number>
    Starting Spark Job = ef3b76c8-2438-4a73-bcca-2f6960c5a01e
    Job hasn't been submitted after 61s. Aborting it.
    Possible reasons include network issues, errors in remote driver or the cluster has no available resources, etc.
    Please check YARN or Spark driver's logs for further information.
    Status: SENT
    10:10:21.625 [16caa78a-7bea-4435-85b7-f0b355dbeca1 main] ERROR org.apache.hadoop.hive.ql.exec.spark.status.SparkJobMonitor - Job hasn't been submitted after 61s. Aborting it.
    Possible reasons include network issues, errors in remote driver or the cluster has no available resources, etc.
    Please check YARN or Spark driver's logs for further information.
    10:10:21.626 [16caa78a-7bea-4435-85b7-f0b355dbeca1 main] ERROR org.apache.hadoop.hive.ql.exec.spark.status.SparkJobMonitor - Status: SENT
    FAILED: Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.spark.SparkTask
    10:10:22.019 [16caa78a-7bea-4435-85b7-f0b355dbeca1 main] ERROR org.apache.hadoop.hive.ql.Driver - FAILED: Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.spark.SparkTask

    Yarn webui端:

    #Hive on Spark模式

    #修改hive-site.xml配置:

    <property> <name>spark.master</name> <value>yarn-client</value> </property>
    修改为

    <property> <name>spark.master</name> <value>spark://192.168.66.66:7077</value> </property>

    #修改spark-default.conf配置:

    spark.master  yarn-cluster  为 spark.master   spark://192.168.66.66:7077

    #重启spark

    hive>select count(*) from test.test;

    #结合Hive客户端信息,打开http://192.168.66.66:18080查看Spark任务状态

    结果:执行成功

    Hive端:

    Starting Spark Job = 1ce13e09-0745-4230-904f-5bcd5e90eebb
    Job hasn't been submitted after 61s. Aborting it.
    Possible reasons include network issues, errors in remote driver or the cluster has no available resources, etc.
    Please check YARN or Spark driver's logs for further information.
    10:18:10.765 [8c59fd25-3aac-445e-b34a-5f62e1729694 main] ERROR org.apache.hadoop.hive.ql.exec.spark.status.SparkJobMonitor - Job hasn't been submitted after 61s. Aborting it.
    Possible reasons include network issues, errors in remote driver or the cluster has no available resources, etc.
    Please check YARN or Spark driver's logs for further information.
    Status: SENT
    10:18:10.773 [8c59fd25-3aac-445e-b34a-5f62e1729694 main] ERROR org.apache.hadoop.hive.ql.exec.spark.status.SparkJobMonitor - Status: SENT
    FAILED: Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.spark.SparkTask
    10:18:10.925 [8c59fd25-3aac-445e-b34a-5f62e1729694 main] ERROR org.apache.hadoop.hive.ql.Driver - FAILED: Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.spark.SparkTask 

    Spark web ui:

    #Spark on Yarn模式

    $ ./bin/spark-submit --class org.apache.spark.examples.SparkPi  --master yarn   --deploy-mode client  --driver-memory 512m  --executor-memory 512m  --executor-cores 1   --queue thequeue   lib/spark-examples-1.6.0-hadoop2.4.0.jar   10

    结果:Pi is roughly 3.140388 

    Yarn webui端:

    #Spark on Spark单例模式

    单例模式:修改spark-default.conf中的spark.master为spark://192.168.66.66:7077  在执行时设置--master spark://192.168.66.66:7077 

    ./bin/spark-submit --class org.apache.spark.examples.SparkPi --master spark://192.168.66.66:7077 --deploy-mode client lib/spark-examples-1.6.0-hadoop2.4.0.jar 10

    结果:Pi is roughly 3.14184 

     备注:hive on spark中没有包括hive编译后的spark不能使用sparksql,因为sparksql需要编译hive的包。

     #hive --service  metastore

    #重新下载包-拷贝hive-site.xml到spark/conf下,加入配置<property>
          <name>hive.metastore.uris</name>
          <value>thrift://192.168.66.66:9083</value>
        </property>

  • 相关阅读:
    [转贴]彻底解决 CrystalReports 登录失败问题。
    [资源]《就说》——《大学自习室》作者郝雨又一力作,继续那种搞笑的说唱
    [推荐]一个用来给控件做提示的JS脚本,鼠标移到控件上时出现提示,离开时消失!
    [原创]一个公告栏的源码(利用marquee作的)
    [转帖]DataGrid显示双层表头,即可实现合并单元格问题
    [转贴]15句让女生爱你一生的情话
    [疑问]您没有调试该服务器的权限。验证您是服务器上"Debuger Users"组的成员。
    [转帖]九个不可不占的便宜
    [转贴]秘芨(男孩子为了终身幸福,最好背下来哦)
    [转贴]也许放弃也是一种爱!++++++++觉得这篇文章写得不错
  • 原文地址:https://www.cnblogs.com/xinfang520/p/7763328.html
Copyright © 2020-2023  润新知