• hive on spark的安装及问题


    配置hive

    hive-site

    <property>
    
       <name>hive.metastore.uris</name>
    
       <value>thrift://database:9083</value>
    
    </property> 
    
    <property>
    
       <name>hive.metastore.client.socket.timeout</name>
    
       <!--<value>600s</value>-->
    
        <value>600</value>
    
    </property>
    

    把hive-site.xml 放到spark/conf目录下

    Mysql驱动放到spark/lib目录下

    启动:hive --service metastore

    配置spark

    Slaves

    spark04
    spark02
    

    Spark-env.sh

    SPARK_MASTER_IP=spark02
    
    JAVA_HOME=/usr/local/jdk1.7.0_75
    
    SPAKR_HIVE=true
    
    HADOOP_CONF_DIR=/usr/local/hadoop-2.6.0/etc/hadoop
    

    spark-defaults.conf

    # Default system properties included when running spark-submit.
    # This is useful for setting default environmental settings.
    
    # Example:
    # spark.master                     spark://master:7077
    spark.eventLog.enabled           true
    #spark.eventLog.dir               hdfs://mycluster:8021/spark/logs/events
    # spark.eventLog.dir               hdfs://namenode:8021/directory
    # spark.serializer                 org.apache.spark.serializer.KryoSerializer
    # spark.driver.memory              5g
    # spark.executor.extraJavaOptions  -XX:+PrintGCDetails -Dkey=value -Dnumbers="one two three"
    

     

    Scp到其它机器上

    测试spark-hive

    spark-shell--master spark://spark02:7077
    
    valsqlContext = new org.apache.spark.sql.hive.HiveContext(sc);
    
    sqlContext.sql("selectcount(*) from ods_app.dev_location").collect().foreach(println);
    

    相关问题

    1、hive metastore 问题

    java.lang.RuntimeException: Unable to instantiateorg.apache.hadoop.hive.metastore.HiveMetaStoreClient
    
    解决方法:
    
    在hive-site.xml配置hive.metastore.uris,并启动hive metastore
    
    <property>
    
       <name>hive.metastore.uris</name>
    
       <value>thrift://database:9083</value>
    
    </property>
    


    2、Ha mycluster的问题

    java.lang.IllegalArgumentException:java.net.UnknownHostException: mycluster
    
    解决方法:
    
    在spark-env.sh,配置HADOOP_CONF_DIR
    
    HADOOP_CONF_DIR=/usr/local/hadoop-2.6.0/etc/hadoop
    
  • 相关阅读:
    (一)lamp 环境搭建之编译安装apache
    ssl 原理和建立连接过程
    lamp 架构的理解
    diff 命令,防止遗忘
    关于云计算的一些概念理解
    lvs 四层负载相关
    Mac 安装Git
    Harbor镜像仓库v2.1.0_2安装与使用
    CentOS部署Harbor镜像仓库
    从harbor部署到在k8s中使用
  • 原文地址:https://www.cnblogs.com/smarthhl/p/5101971.html
Copyright © 2020-2023  润新知