• phoenix 利用CsvBulkLoadTool 批量带入数据并自动创建索引


    需要先创建表:

    CREATE TABLE IF NOT EXISTS population (
    state CHAR(2) NOT NULL, city VARCHAR NOT NULL, population BIGINT
    CONSTRAINT my_pk PRIMARY KEY (state, city));

    在phoenix 目录下执行

    hadoop jar /home/phoenix-4.6.0-HBase-1.0-bin/phoenix-4.6.0-HBase-1.0-client.jar   org.apache.phoenix.mapreduce.CsvBulkLoadTool -t POPULATION -i /datas/us_population.csv

    -t :tableName

    -i: input file 文件必须在hdfs文件上。

    后查询表数据是空。

    问题: ERROR mapreduce.CsvBulkLoadTool: Error Wrong FS: file:/home/hadoop/tmp/partitions_101bd67a-ec2c-4808-bc9f-bf4cd6ea74b9, expected: hdfs://node11:9000 occurred submitting CSVBulkLoad 

     参数意义:

    -a,--array-delimiter <arg> Array element delimiter (optional)
    -c,--import-columns <arg> Comma-separated list of columns to be
    imported
    -d,--delimiter <arg> Input delimiter, defaults to comma
    -e,--escape <arg> Supply a custom escape character, default is
    a backslash
    -g,--ignore-errors Ignore input errors
    -h,--help Show this help and quit
    -i,--input <arg> Input CSV path (mandatory)
    -it,--index-table <arg> Phoenix index table name when just loading
    this particualar index table
    -o,--output <arg> Output path for temporary HFiles (optional)
    -q,--quote <arg> Supply a custom phrase delimiter, defaults
    to double quote character
    -s,--schema <arg> Phoenix schema name (optional)
    -t,--table <arg> Phoenix table name (mandatory)
    -z,--zookeeper <arg> Supply zookeeper connection details
    (optional)

  • 相关阅读:
    在linux下如何判断是否已经安装某个软件?
    $ cd `dirname $0` 和PWD用法
    linux下添加,删除,修改,查看用户和用户组
    客户端远程连接linux下mysql数据库授权
    MySQL各个版本区别
    查看linux系统类型、版本、位数
    /bin/bash^M: bad interpreter: No such file or directory
    npm note
    karma note
    jasmine note
  • 原文地址:https://www.cnblogs.com/zhanggl/p/5191912.html
Copyright © 2020-2023  润新知