• 【TPCDS】trino+S3+hive+postgresql性能测试hadoop+mysql+hive安装部署(三)


    总体过程

    1、安装hadoop-3.2.0

    2、安装hive-standalone-metastore-3.0.0

    3、配置环境变量

     

    安装hadoop-3.2.0

    tar -zxvf hadoop-3.2.0.tar.gz -C /opt/
    • 重定向
    ln -sv /opt/hadoop-3.2.0 /opt/hadoop
    ln -sv /opt/hadoop/share/hadoop/tools/lib/hadoop-aws* /opt/hadoop/share/hadoop/common/lib/
    ln -sv /opt/hadoop/share/hadoop/tools/lib/aws-java-sdk* /opt/hadoop/share/hadoop/common/lib/
    • 配置hadoop环境
    vim /etc/profile

    添加如下的内容

    #set java environment
    export JAVA_HOME CLASSPATH PATH
    export HADOOP_HOME=/opt/hadoop
    export PATH=$PATH:$HADOOP_HOME/bin

     

    安装hive-standalone-metastore-3.0.0

    tar -zxvf hive-standalone-metastore-3.0.0-bin.tar.gz -C /opt/
    • 重定向
    ln -s /opt/apache-hive-metastore-3.0.0-bin /opt/hive-metastore
    sed -i 's#/dev/null#/tmp/metastore_start.log#' /opt/apache-hive-metastore-3.0.0-bin/bin/start-metastore
    sed -i 's/256/1024/g' /opt/hive-metastore/bin/metastore-config.sh

    安装mysql-connector-java-5.1.47

    tar -zxvf mysql-connector-java-5.1.47 /opt/
    • 重定向
    ln -s /opt/mysql-connector-java-5.1.47/mysql-connector-java-5.1.47.jar /opt/hadoop/share/hadoop/common/lib/ 
    ln -s /opt/mysql-connector-java-5.1.47/mysql-connector-java-5.1.47.jar /opt/hive-metastore/lib/

     

    修改和添加配置文件metastore-site.xml

    • 进入目录/opt/hive-metastore/conf/,将hive-site.xml改为metastore-site.xml。将如下的内容复制替换掉metastore-site.xml的文件中
    <?xml version="1.0"?>
    <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
    <!--
       Licensed to the Apache Software Foundation (ASF) under one or more
       contributor license agreements.  See the NOTICE file distributed with
       this work for additional information regarding copyright ownership.
       The ASF licenses this file to You under the Apache License, Version 2.0
       (the "License"); you may not use this file except in compliance with
       the License.  You may obtain a copy of the License at
    
           http://www.apache.org/licenses/LICENSE-2.0
    
       Unless required by applicable law or agreed to in writing, software
       distributed under the License is distributed on an "AS IS" BASIS,
       WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
       See the License for the specific language governing permissions and
       limitations under the License.
    -->
    
    <configuration>
    <property>
            <name>metastore.task.threads.always</name>
            <value>org.apache.hadoop.hive.metastore.events.EventCleanerTask</value>
    </property>
    <property>
            <name>metastore.expression.proxy</name>
            <value>org.apache.hadoop.hive.metastore.DefaultPartitionExpressionProxy</value>
    </property>
    
    <property>
      <name>hive.metastore.warehouse.dir</name>
      <value>s3a://fastdata-tsdb/syw/warehouse/</value>
    </property>
    
    <property>
      <name>hive.metastore.local</name>
      <value>false</value>
    </property>
     
    <property>
      <name>hive.metastore.uris</name>
      <value>thrift://10.201.0.124:9083</value>
    </property>
    
    <property>
      <name>fs.s3a.fast.upload</name>
      <value>true</value>
    </property>
    <property>
      <name>fs.s3a.multipart.size</name>
      <value>100M</value>
    </property>
     <property>
       <name>hive.metastore.local</name>
         <value>false</value>
    </property>
     <property>
       <name>hive.metastore.schema.verification</name>
         <value>false</value>
    </property>
    <property>
            <name>javax.jdo.option.ConnectionDriverName</name>
            <value>com.mysql.jdbc.Driver</value>
    </property>
    <property>
            <name>javax.jdo.option.ConnectionURL</name>
            <value>jdbc:mysql://10.201.0.125:3306/metastore_db?useSSL=false</value>
    </property>
    <property>
            <name>javax.jdo.option.ConnectionUserName</name>
            <value>root</value>
    </property>
    <property>
            <name>javax.jdo.option.ConnectionPassword</name>
            <value>123456</value>
    </property>
    
    </configuration>

    【重要】如下的内容字段需要根据自己的情况进行修改,其他的内容不变:

    • hive.metastore.uris:自己的服务器地址,注意修改10.201.0.124替换成自己的服务器IP
    • javax.jdo.option.ConnectionURL:自己安装的mysql的连接地址
    • javax.jdo.option.ConnectionUserName:自己mysql数据库的用户名称
    • javax.jdo.option.ConnectionPassword:自己mysql数据库的密码

    修改和添加配置文件core-site.xml

    • 进入cd /opt/hadoop/etc/hadoop/ 修改core-site.xml文件,用如下的代码内容替换掉原先的内容即可,不需要做任何的修改
    <?xml version="1.0"?>
    <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
    <configuration>
    <property> 
       <name>fs.s3a.impl</name> 
       <value>org.apache.hadoop.fs.s3a.S3AFileSystem</value> 
    </property> 
    <property> 
       <name>fs.s3a.access.key</name> 
       <value>GVO55HIKELKQIN4QSOPL</value> 
    </property> 
    <property> 
       <name>fs.s3a.secret.key</name> 
       <value>XgdhYuGsXHHg3yi0WySVQ8GwztW4SpIrL5irUdlA</value> 
    </property> 
    <property> 
       <name>fs.s3a.endpoint</name> 
       <value>obs.cn-southwest-2.myhuaweicloud.com</value> 
    </property> 
    <property>
    <name>fs.s3a.list.version</name>
    <value>1</value>
    </property>
    </configuration>

     

    catalog中添加hive、postgres、iceberg配置文件

    注意:每个服务器都需要添加,直接复制粘贴即可,不需要做任何的修改

    • 进入目录:/root/trino/trino-server-363/etc/catalog
    • 新建hive.properties、postgresql.properties、iceberg.properties文件
    • hive.properties。10.201.0.124注意修改为master的IP,其余内容均不修改
    connector.name=hive-hadoop2
    hive.metastore.uri=thrift://10.201.0.124:9083
    hive.allow-drop-table=true
    hive.max-partitions-per-writers=10000
    hive.max-partitions-per-scan=1000000
    hive.s3.endpoint=obs.cn-southwest-2.myhuaweicloud.com
    hive.s3.path-style-access=true
    hive.s3.ssl.enabled=false
    hive.s3.max-connections=100

    postgresql.properties。(注意:tpcds是pq数据中常见的库名)

    connector.name=postgresql
    connection-url=jdbc:postgresql://10.201.0.125:5432/tpcds
    connection-user=postgres
    connection-password=public
    • iceberg.properties,10.201.0.124注意修改为master的IP,其余内容均不修改
    connector.name=iceberg
    hive.metastore.uri=thrift://10.201.0.124:9083
    hive.max-partitions-per-scan=1000000
    hive.s3.endpoint=obs.cn-southwest-2.myhuaweicloud.com
    hive.s3.path-style-access=true
    hive.s3.ssl.enabled=false
    hive.s3.max-connections=100

    启动metastore(124)

    /opt/hive-metastore/bin/start-metastore -p 9083

    创建metastore_db库

    • 连接mysql,并手动创建库,库名称为:metastore_db
    mysql -uroot -h 10.201.0.125 -p
    输入密码
    create database metastore_db;

    mysql库metastore_db中创建表

    /opt/hive-metastore/bin/schematool --verbose -initSchema -dbType mysql -userName root  -passWord 123456 -url jdbc:mysql://10.201.0.125:3306/metastore_db?useSSL=false

    【重要】注意修改10.201.0.125为自己配置的mysql服务器地址

    • 进入mysql中验证库中的表

    启动trino

    注:若之前已启动trino。则此处不用再次启动

    验证trino环境是否ok

    验证环境

    ./trino --server 10.201.0.125:8080 --catalog postgresql

    1.作者:Syw
    2.出处:http://www.cnblogs.com/syw20170419/
    3.本文版权归作者和博客园共有,欢迎转载,但未经作者同意必须保留此段声明,且在文章页面明显位置给出原文连接,否则保留追究法律责任的权利。
    4.如果文中有什么错误,欢迎指出。以免更多的人被误导。
  • 相关阅读:
    Centos7更改网卡名为eth0
    Centos7部署Open-Falcon监控
    centos6.x一键15项系统优化(转自努力哥)
    运维题目(十三)
    运维题目(十二)
    Mongodb的学习整理(下)
    Centos7下yum安装mongodb
    浏览器缓存
    控制反转
    js setTimeOut()
  • 原文地址:https://www.cnblogs.com/syw20170419/p/15593357.html
Copyright © 2020-2023  润新知