• springboot+kafka+sparkstreaming 生产及消费数据-超简单实例


    springboot+kafka+sparkstreaming 生产及消费数据-超简单实例
    kafka生产者实例:
    import org.apache.kafka.clients.producer.Callback;
    import org.apache.kafka.clients.producer.KafkaProducer;
    import org.apache.kafka.clients.producer.ProducerRecord;
    import org.apache.kafka.clients.producer.RecordMetadata;
    import org.slf4j.Logger;
    import org.slf4j.LoggerFactory;

    import java.io.IOException;
    import java.io.InputStream;
    import java.util.Properties;
    import java.util.Random;

    public class LiveServerLog {
    private static final Logger LOGGER = LoggerFactory.getLogger(LiveServerLog.class);
    private int retry;
    private static KafkaProducer<String, String> kafkaProducer;
    private static final LiveServerLog INSTANCE = new LiveServerLog();
    private LiveServerLog() {
    }
    public static final LiveServerLog getInstance() {
    return INSTANCE;
    }

    /**
    * kafka生产者进行初始化
    * @param retry 重试次数
    */
    public void initConfig(int retry) {
    this.retry = retry;
    if (null == kafkaProducer) {
    Properties props = new Properties();
    InputStream inStream = null;
    try {
    inStream = this.getClass().getClassLoader()
    .getResourceAsStream("kafka.properties");
    props.load(inStream);
    kafkaProducer = new KafkaProducer<String, String>(props);
    } catch (IOException e) {
    LOGGER.error("kafkaProducer初始化失败:" + e.getMessage(), e);
    } finally {
    if (null != inStream) {
    try {
    inStream.close();
    } catch (IOException e) {
    LOGGER.error("kafkaProducer初始化失败:" + e.getMessage(), e);
    }
    }
    }
    }
    }

    /**
    * 通过kafkaProducer发送消息
    * @param topic 消息接收主题
    * @param message 具体消息值
    */
    public void sendKafkaMessage(String topic, String message) {
    /**
    * 1、如果指定了某个分区,会只讲消息发到这个分区上 random.nextInt(2)
    * 2、如果同时指定了某个分区和key,则也会将消息发送到指定分区上,key不起作用 random.nextInt(2), "",
    * 3、如果没有指定分区和key,那么将会随机发送到topic的分区中
    * 4、如果指定了key,那么将会以hash<key>的方式发送到分区中
    */
    ProducerRecord<String, String> record = new ProducerRecord<String, String>(
    topic, message);
    // send方法是异步的,添加消息到缓存区等待发送,并立即返回,这使生产者通过批量发送消息来提高效率
    // kafka生产者是线程安全的,可以单实例发送消息
    kafkaProducer.send(record, new Callback() {
    @Override
    public void onCompletion(RecordMetadata metadata, Exception exception) {
    if (exception != null) {
    exception.printStackTrace();
    retryKakfaMessage(topic, message);
    } else {
    System.out.println(metadata.topic() + "-" + metadata.partition());
    }
    }
    });
    }

    /**
    * 当kafka消息发送失败后,重试
    */
    private void retryKakfaMessage(String topic, String retryMessage) {
    ProducerRecord<String, String> record = new ProducerRecord<String, String>(
    topic, retryMessage);
    for (int i = 1; i <= retry; i++) {
    try {
    kafkaProducer.send(record);
    return;
    } catch (Exception e) {
    LOGGER.error("kafka发送消息失败:" + e.getMessage(), e);
    retryKakfaMessage(topic, retryMessage);
    }
    }
    }
    }
    kafka.properties
    bootstrap.servers=10.105.1.4:9092,10.105.1.5:9092,10.105.1.6:9092
    acks=1
    retries=3
    batch.size=1000
    key.serializer=org.apache.kafka.common.serialization.StringSerializer
    value.serializer=org.apache.kafka.common.serialization.StringSerializer
    client.id=producer.Live_Server.Log
    springboot调用实例:
    import net.sf.json.JSONObject;
    import org.slf4j.Logger;
    import org.slf4j.LoggerFactory;
    import org.springframework.boot.SpringApplication;
    import org.springframework.boot.autoconfigure.SpringBootApplication;
    import org.springframework.context.annotation.Import;
    import org.springframework.scheduling.annotation.EnableScheduling;
    import org.springframework.transaction.annotation.EnableTransactionManagement;
    import org.springframework.web.bind.annotation.RequestMapping;
    import org.springframework.web.bind.annotation.ResponseBody;
    import org.springframework.web.bind.annotation.RestController;

    import javax.servlet.http.HttpServletRequest;
    import javax.servlet.http.HttpServletResponse;

    @RestController
    @SpringBootApplication
    @EnableTransactionManagement
    @EnableScheduling
    @RequestMapping(value = "/LiveService/*")
    public class LiveService {
    private final static Logger log = LoggerFactory.getLogger(LiveService.class);

    public static void main(String[] args) throws Exception {
    SpringApplication.run(LiveService.class, args);
    }

    @RequestMapping(value = "/", produces = {"text/plain;charset=UTF-8"})
    @ResponseBody
    public String returnString() {
    return "Hello LiveService";
    }

    /**
    * 记录日志
    */
    @RequestMapping(value = "LiveServerLog", produces = {"application/json;charset=UTF-8"})
    @ResponseBody
    public void LiveServerLog(HttpServletRequest request, HttpServletResponse response) {
    try {
    JSONObject _condition = getStringFromStream(request);
    String log = _condition.getString("log");
    LiveServerLog.getInstance().initConfig(3);
    LiveServerLog.getInstance().sendKafkaMessage("live_server_log", log);
    } catch (Exception e) {
    log.info(e.getMessage());
    }
    }

    /**
    * 获取请求参数
    */
    private JSONObject getStringFromStream(HttpServletRequest req) {
    ServletInputStream is;
    try {
    is = req.getInputStream();
    int nRead = 1;
    int nTotalRead = 0;
    byte[] bytes = new byte[102400];
    while (nRead > 0) {
    nRead = is.read(bytes, nTotalRead, bytes.length - nTotalRead);
    if (nRead > 0)
    nTotalRead = nTotalRead + nRead;
    }
    String str = new String(bytes, 0, nTotalRead, "utf-8");
    return JSONObject.fromObject(str);
    } catch (IOException e) {
    e.printStackTrace();
    return null;
    }
    }
    }
    sparkstreaming消费数据:
    import org.apache.kafka.common.serialization.StringDeserializer
    import org.apache.spark.SparkConf
    import org.apache.spark.sql.{Row, SparkSession}
    import org.apache.spark.streaming.{Seconds, StreamingContext}
    import org.apache.spark.streaming.kafka010.ConsumerStrategies.Subscribe
    import org.apache.spark.streaming.kafka010.{CanCommitOffsets, HasOffsetRanges, KafkaUtils}
    import org.apache.spark.streaming.kafka010.LocationStrategies.PreferConsistent
    import org.slf4j.LoggerFactory

    import java.util.ResourceBundle

    class live_server_log

    object live_server_log {
    private val LOGGER = LoggerFactory.getLogger(classOf[live_server_log])

    def main(args: Array[String]): Unit = {
    try {
    val conf = new SparkConf().setAppName("live_server_log").setMaster("yarn-cluster") //.setMaster("local")//
    // spark2用法
    val ss = SparkSession.builder.config(conf).getOrCreate()
    val ssc = new StreamingContext(ss.sparkContext, Seconds.apply(5))//5秒执行一次
    val prop = ResourceBundle.getBundle("app")
    val bootstrapServers = prop.getString("bootstrap.servers")
    val kafkaParams = Map[String, Object](
    "bootstrap.servers" -> bootstrapServers, // kafka 集群
    "key.deserializer" -> classOf[StringDeserializer],
    "value.deserializer" -> classOf[StringDeserializer],
    "group.id" -> "Kafka Broker Default Group",
    "auto.offset.reset" -> "earliest", // 每次都是从头开始消费(from-beginning),可配置其他消费方式
    "enable.auto.commit" -> (false: java.lang.Boolean) //手动提交偏移量
    )
    val topics = Array("live_server_log") //主题,可配置多个
    val stream = KafkaUtils.createDirectStream[String, String](
    ssc,
    PreferConsistent,
    Subscribe[String, String](topics, kafkaParams)
    )
    // val list = List(
    // StructField("S_USER_NAME", StringType, nullable = true),
    // StructField("D_CREATE", StringType, nullable = true)
    // )
    // val schema = StructType(list)
    stream.foreachRDD(rdd => {
    //计算偏移量
    val offsetRanges = rdd.asInstanceOf[HasOffsetRanges].offsetRanges
    if (!rdd.isEmpty()) rdd.foreach(line => {
    println(line.value())
    //自行选择入库方式
    //insertLog(line.value())
    // val log_rdd1 = rdd.map(r => {
    // createRow(r.value().toString())
    // })
    // val dataFrame = ss.createDataFrame(log_rdd1, schema)
    // val date = Common.getToday(Common.DateFormat.targetDAY.getValue)
    // dataFrame.write.format("parquet").mode(SaveMode.Append).save("hdfs://10.105.1.1:8020/user/hive/warehouse/default_db/log/" + date)
    })
    //设定偏移量
    stream.asInstanceOf[CanCommitOffsets].commitAsync(offsetRanges)
    })
    ssc.start()
    ssc.awaitTermination()
    // ...
    } catch {
    case ex: Exception => {
    ex.printStackTrace() // 打印到标准err
    LOGGER.error("kafka消费消息失败:" + ex.getMessage, ex)
    }
    }
    }

    /**
    * 创建行数据
    * @return Row
    */
    def createRow(s: String): Row = {
    val l = s.split(",")
    val row = Row(l(0), l(1), l(2))
    return row
    }

    /**
    * 入库
    */
    def insertLog(s: String) {
    if (!s.trim().isEmpty) {
    val l = s.split(",")
    //call_oracle.getInstance().callLiveClientLog(l(0), l(1), l(2).toInt)
    }
    }
    }
    提交到spark2执行:
    spark2-submit --master yarn --deploy-mode cluster --driver-memory 1g --executor-memory 1g /home/kafka/live_server_log.jar --class com.kafka.live_server_log
    pom1.xml
    <parent>
    <groupId>org.springframework.boot</groupId>
    <artifactId>spring-boot-starter-parent</artifactId>
    <version>1.5.9.RELEASE</version>
    <relativePath/> <!-- lookup parent from repository -->
    </parent>
    <properties>
    <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
    <maven.compiler.source>1.8</maven.compiler.source>
    <maven.compiler.target>1.8</maven.compiler.target>
    <logback.version>1.2.3</logback.version>
    <scala.maven.version>3.2.0</scala.maven.version>
    <scala.binary.version>2.10.5</scala.binary.version>
    <scala.version>2.10.5</scala.version>
    <spark.version>1.6.0-cdh5.14.0</spark.version>
    </properties>

    <dependencies>
    <!-- jdbc -->
    <dependency>
    <groupId>org.springframework.boot</groupId>
    <artifactId>spring-boot-starter-jdbc</artifactId>
    </dependency>
    <!-- web -->
    <dependency>
    <groupId>org.springframework.boot</groupId>
    <artifactId>spring-boot-starter-web</artifactId>
    </dependency>
    <dependency>
    <groupId>junit</groupId>
    <artifactId>junit</artifactId>
    <version>4.11</version>
    <scope>test</scope>
    </dependency>
    <dependency>
    <groupId>ch.qos.logback</groupId>
    <artifactId>logback-core</artifactId>
    <version>${logback.version}</version>
    </dependency>
    <dependency>
    <groupId>ch.qos.logback</groupId>
    <artifactId>logback-classic</artifactId>
    <version>${logback.version}</version>
    </dependency>
    <dependency>
    <groupId>org.apache.commons</groupId>
    <artifactId>commons-lang3</artifactId>
    <version>3.7</version>
    </dependency>
    <dependency>
    <groupId>commons-io</groupId>
    <artifactId>commons-io</artifactId>
    <version>2.6</version>
    </dependency>
    <dependency>
    <groupId>org.apache.hadoop</groupId>
    <artifactId>hadoop-client</artifactId>
    <version>2.6.0-cdh5.14.0</version>
    <exclusions>
    <exclusion>
    <groupId>log4j</groupId>
    <artifactId>log4j</artifactId>
    </exclusion>
    <exclusion>
    <artifactId>slf4j-log4j12</artifactId>
    <groupId>org.slf4j</groupId>
    </exclusion>
    </exclusions>
    </dependency>
    <dependency>
    <groupId>org.apache.hadoop</groupId>
    <artifactId>hadoop-common</artifactId>
    <version>2.6.0-cdh5.14.0</version>
    <exclusions>
    <exclusion>
    <groupId>log4j</groupId>
    <artifactId>log4j</artifactId>
    </exclusion>
    <exclusion>
    <artifactId>slf4j-log4j12</artifactId>
    <groupId>org.slf4j</groupId>
    </exclusion>
    <exclusion>
    <artifactId>servlet-api</artifactId>
    <groupId>javax.servlet</groupId>
    </exclusion>
    </exclusions>
    </dependency>
    <dependency>
    <groupId>commons-logging</groupId>
    <artifactId>commons-logging</artifactId>
    <version>1.2</version>
    </dependency>
    <!-- hadoop mr -->
    <dependency>
    <groupId>org.apache.hadoop</groupId>
    <artifactId>hadoop-mapreduce-client-core</artifactId>
    <version>2.6.0-cdh5.14.0</version>
    <exclusions>
    <exclusion>
    <groupId>log4j</groupId>
    <artifactId>log4j</artifactId>
    </exclusion>
    <exclusion>
    <artifactId>slf4j-log4j12</artifactId>
    <groupId>org.slf4j</groupId>
    </exclusion>
    <exclusion>
    <artifactId>servlet-api</artifactId>
    <groupId>javax.servlet</groupId>
    </exclusion>
    </exclusions>
    </dependency>
    <!-- parquet列式存储 -->
    <dependency>
    <groupId>org.apache.parquet</groupId>
    <artifactId>parquet-hadoop</artifactId>
    <version>1.8.1</version>
    </dependency>
    <!-- json格式化 -->
    <dependency>
    <groupId>net.sf.json-lib</groupId>
    <artifactId>json-lib</artifactId>
    <version>2.4</version>
    </dependency>
    <!-- https://mvnrepository.com/artifact/org.apache.kafka/kafka -->
    <dependency>
    <groupId>org.apache.kafka</groupId>
    <artifactId>kafka_2.10</artifactId>
    <version>0.10.2.2</version>
    <exclusions>
    <exclusion>
    <groupId>org.slf4j</groupId>
    <artifactId>slf4j-log4j12</artifactId>
    </exclusion>
    </exclusions>
    </dependency>
    </dependencies>
    pom2.xml
    <dependency>
    <groupId>org.apache.spark</groupId>
    <artifactId>spark-sql_2.11</artifactId>
    <version>2.4.0.cloudera2</version>
    <exclusions>
    <exclusion>
    <groupId>log4j</groupId>
    <artifactId>log4j</artifactId>
    </exclusion>
    <exclusion>
    <artifactId>slf4j-log4j12</artifactId>
    <groupId>org.slf4j</groupId>
    </exclusion>
    </exclusions>
    </dependency>
    <dependency>
    <groupId>org.apache.spark</groupId>
    <artifactId>spark-streaming-kafka-0-10_2.11</artifactId>
    <version>2.4.0.cloudera2</version>
    <exclusions>
    <exclusion>
    <groupId>log4j</groupId>
    <artifactId>log4j</artifactId>
    </exclusion>
    <exclusion>
    <artifactId>slf4j-log4j12</artifactId>
    <groupId>org.slf4j</groupId>
    </exclusion>
    <exclusion>
    <artifactId>slf4j-api</artifactId>
    <groupId>org.slf4j</groupId>
    </exclusion>
    </exclusions>
    </dependency>
    <dependency>
    <groupId>org.apache.spark</groupId>
    <artifactId>spark-streaming_2.11</artifactId>
    <version>2.4.0.cloudera2</version>
    <exclusions>
    <exclusion>
    <groupId>log4j</groupId>
    <artifactId>log4j</artifactId>
    </exclusion>
    <exclusion>
    <artifactId>slf4j-log4j12</artifactId>
    <groupId>org.slf4j</groupId>
    </exclusion>
    <exclusion>
    <artifactId>slf4j-api</artifactId>
    <groupId>org.slf4j</groupId>
    </exclusion>
    </exclusions>
    </dependency>
    <dependency>
    <groupId>log4j</groupId>
    <artifactId>log4j</artifactId>
    <version>1.2.17</version>
    </dependency>
    ————————————————
    版权声明:本文为CSDN博主「史诗级大菠萝」的原创文章,遵循CC 4.0 BY-SA版权协议,转载请附上原文出处链接及本声明。
    原文链接:https://blog.csdn.net/weixin_43827665/article/details/116052515

  • 相关阅读:
    迁移模型问题,提示admin已存在
    centos 部署的时候安装不上Mariadb,缺少依赖文件
    collections
    List里面添加子list,子list clear之后竟然会影响主List里面的内容
    Jackson用法详解
    Ouath2.0在SpringCloud下验证获取授权码
    zookeeper原理之Leader选举的getView 的解析流程和ZkServer服务启动的逻辑
    zookeeper原理之Leader选举源码分析
    Spring Integration sftp 专栏详解
    SpringMVC常用注解标签详解
  • 原文地址:https://www.cnblogs.com/javalinux/p/15060151.html
Copyright © 2020-2023  润新知