• 集群资源队列监控:Grafana


    集群资源队列监控:Grafana

    common.sh

    #!/bin/sh
    
    . ~/.bashrc
    
    home=$(cd `dirname $0`; cd ..; pwd)
    bin_home=$home/bin
    config_home=$home/conf
    logs_home=$home/logs
    lib_home=$home/lib
    data_home=$home/data
    
    cluster=$1
    day=$2
    
    if [ "${cluster}" = "" -o "${day}" = "" ]; then
        echo "请指定参数集群和日期(例:yth 20191203):"
        exit 0
    fi
    
    day_timestamp=`date -d "${day}" +%s`
    day_timestamp_1=`date -d "${day} +1 day" +%s`
    day_str0=`date -d @${day_timestamp} "+%Y-%m-%d"`
    day_str1=`date -d @${day_timestamp} "+%Y/%m/%d"`
    day_str2=`date -d @${day_timestamp_1} "+%Y-%m-%d"`job_parse_input_path
    
    conf=${config_home}/${cluster}.conf
    
    flink_home=`cat ${conf} | grep ^flink_home | awk -F"=" '{print $2}'`
    
    fsimage_binary_path=`cat ${conf} | grep ^fsimage_binary_path | awk -F"=" '{print $2}'`
    fsimage_txt_path=`cat ${conf} | grep ^fsimage_txt_path | awk -F"=" '{print $2}'`
    fsimage_org_hdfs_path=`cat ${conf} | grep ^fsimage_org_hdfs_path | awk -F"=" '{print $2}'`
    fsimage_detail_hdfs_path=`cat ${conf} | grep ^fsimage_detail_hdfs_path | awk -F"=" '{print $2}'`
    
    fsimage_org_hdfs_path=${fsimage_org_hdfs_path}/day=${day}/cluster=${cluster}
    fsimage_detail_hdfs_path=${fsimage_detail_hdfs_path}/day=${day}/cluster=${cluster}
    
    yarn_queue_offline=`cat ${conf} | grep ^yarn_queue_offline | awk -F"=" '{print $2}'`
    
    #作业解析参数
    job_parse_input_path=`cat ${conf} | grep ^job_parse_input_path | awk -F"=" '{print $2}'`
    job_parse_output_path=`cat ${conf} | grep ^job_parse_output_path | awk -F"=" '{print $2}'`

    start-fsimage-detail.sh

    #!/bin/sh
    
    home=$(cd `dirname $0`; cd ..; pwd)
    . ${home}/bin/common.sh
    
    hdfs dfs -mkdir -p ${fsimage_detail_hdfs_path}
    hdfs dfs -rm -r ${fsimage_detail_hdfs_path}
    
    hadoop jar ${lib_home}/ops-fsimage-detail-2.0-release.jar \
    com.djt.fsimage.detail.mrdetail.FsImageLogMR \
    -Dmapreduce.job.queuename=${yarn_queue_offline} \
    -DinputPath=${fsimage_org_hdfs_path} \
    -DoutputPath=${fsimage_detail_hdfs_path} \
    -Dday=${day} \
    -DreduceCount=20
    
    if [ "${hive_on}" == "true" ]
    then
        hive -e "alter table xunshan.dwd_fsimage_detail add partition (day=${day}, cluster='${cluster}') location '${fsimage_detail_hdfs_path}'"
    fi

    start-fsimage-oiv.sh

    #!/bin/sh
    
    home=$(cd `dirname $0`; cd ..; pwd)
    . ${home}/bin/common.sh
    
    export HADOOP_HEAPSIZE=20000
    
    fsimage_binary_name=`ls ${fsimage_binary_path} | grep ${cluster} | grep ${day}`
    
    fsimage_binary_file=${fsimage_binary_path}/${fsimage_binary_name}
    
    fsimage_txt_name=${fsimage_binary_name}.txt
    fsimage_txt_file=${fsimage_txt_path}/${fsimage_txt_name}
    
    hdfs oiv -p Delimited -i ${fsimage_binary_file} -o ${fsimage_txt_file}
    
    hdfs dfs -mkdir -p ${fsimage_org_hdfs_path}
    hdfs dfs -rm ${fsimage_org_hdfs_path}/${fsimage_txt_name}
    hdfs dfs -put ${fsimage_txt_file} ${fsimage_org_hdfs_path}
    
    echo -e "==========> load to hive start ==========>"
    hive -e "ALTER TABLE xunshan.dwd_fsimage_org DROP IF EXISTS PARTITION(day=${day},cluster='${cluster}');"
    hive -e "ALTER TABLE xunshan.dwd_fsimage_org ADD PARTITION(day=${day},cluster='${cluster}') LOCATION '${fsimage_org_hdfs_path}';"
    echo -e "==========> load to hive end ==========>"

    start-job-parse.sh

    #!/bin/sh
    
    home=$(cd `dirname $0`; cd ..; pwd)
    . ${home}/bin/common.sh
    
    ${flink_home}/bin/flink run -c com.djt.job.parse.JobMain -m yarn-cluster -yjm 8192  -yn 10 -ys 4 -p 40 -ytm 8192 -yqu ${yarn_queue_offline} \
    ${lib_home}/ops-job-parse-2.0-release.jar \
    --cluster ${cluster} \
    --day ${day} \
    --job.input.path ${job_parse_input_path} \
    --job.output.path ${job_parse_output_path}

    start-yarn-monitor.sh

    #!/bin/sh
    
    home=$(cd `dirname $0`; cd ..; pwd)
    . ${home}/bin/common.sh
    
    ${flink_home}/bin/flink run -c com.djt.yarn.monitor.Main -m yarn-cluster -yjm 1024  -yn 1 -ys 1 -p 1 -ytm 1024 -yqu root.realtime \
    ${lib_home}/ops-yarn-monitor-2.0-release.jar \
    --clusterNames cluster001 \
    --cluster001.hosts ops2:8088,ops2:8088 \
    --influxdb.url http://work1:9092 \
    --influxdb.username admin \
    --influxdb.password admin \
    --influxdb.database yarn
    ########## 今天的苦逼是为了不这样一直苦逼下去!##########
  • 相关阅读:
    nodejs MYSQL数据库执行多表查询
    【BZOJ3994】[SDOI2015]约数个数和 莫比乌斯反演
    【BZOJ2693】jzptab 莫比乌斯反演
    【BZOJ2154】Crash的数字表格 莫比乌斯反演
    【BZOJ2242】[SDOI2011]计算器 BSGS
    【BZOJ2005】[Noi2010]能量采集 欧拉函数
    【BZOJ1408】[Noi2002]Robot DP+数学
    【BZOJ2045】双亲数 莫比乌斯反演
    【BZOJ2186】[Sdoi2008]沙拉公主的困惑 线性筛素数
    【BZOJ4176】Lucas的数论 莫比乌斯反演
  • 原文地址:https://www.cnblogs.com/ruii/p/14613487.html
Copyright © 2020-2023  润新知