kafka学习笔记四Consumer学习

1.Consumer and Consumer Group  消费者与消费者分组group.id  分组属性client.id  consumer的id2.Kafka Consumer programming  编程codingpackage com.hanwan.kafka.demo2;import org.apache.kafka.clients.consumer.Co...
摘要由CSDN通过智能技术生成

1.Consumer and Consumer Group  消费者与消费者分组

group.id  分组属性

client.id  consumer的id

2.Kafka Consumer programming  编程coding

package com.hanwan.kafka.demo2;


import org.apache.kafka.clients.consumer.Consumer;
import org.apache.kafka.clients.consumer.ConsumerRecord;
import org.apache.kafka.clients.consumer.ConsumerRecords;
import org.apache.kafka.clients.consumer.KafkaConsumer;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

import java.util.Collection;
import java.util.Collections;
import java.util.Properties;
import java.util.concurrent.atomic.AtomicInteger;

/**
 * @ClassName SimpleConsumer
 * @Description Consumer config demo
 * @Copyright: Copyright (c) 2018</p>
 * @Company: www.lowan.com</ p>
 * @Author hanwan
 * @Date 2018/8/17 16:24
 * @Version 1.0
 **/
public class SimpleConsumer {

    private final static Logger LOGGER = LoggerFactory.getLogger(SimpleConsumer.class);

    public static void main(String[] args) {
        /*
        KafkaConsumer<String, String> consumer = new KafkaConsumer<>(loadProps());
        consumer.subscribe(Collections.singleton("test_c"));
        //表示永久循环
        for (;;) {
            ConsumerRecords<String, String> records = consumer.poll(100);
            records.forEach(record ->{
                LOGGER.info("-----------------------------------------------------------------");
                LOGGER.info("offset {}",record.offset());
                LOGGER.info("value {}", record.value());
                LOGGER.info("partition {}", record.partition());
                LOGGER.info("key {}",record.key());
                LOGGER.info("-----------------------------------------------------------------");
            });
        }*/

        KafkaConsumer<String, String> consumer = new KafkaConsumer<>(loadProps2());
        consumer.subscribe(Collections.singleton("test_c"));
        final AtomicInteger counter = new AtomicInteger();
        while (true) {
            ConsumerRecords<String, String> records = consumer.poll(100);
            records.forEach(record ->{
                LOGGER.info("--------------------------------------------------------");
                LOGGER.info("offset {}", record.offset());
                LOGGER.info("value {}", record.value());
                LOGGER.info("key {}", record.key());
                LOGGER.info("partition {}", record.partition());
                LOGGER.info("--------------------------------------------------------");
                int cnt = counter.getAndIncrement();
                if (cnt >= 3) {
                    Runtime.getRuntime().halt(-1);
                }
            });
        }
    }

    private static Properties loadProps(){
        final Properties prop = new Properties();
        prop.put("bootstrap.servers", "120.55.125.58:9092,120.26.198.248:9092,121.40.200.37:9092");
        prop.put("key.deserializer", "org.apache.kafka.common.serialization.StringDeserializer");
        prop.put("value.deserializer", "org.apache.kafka.common.serialization.StringDeserializer");
        prop.put("group.id", "test_group3");
        prop.put("client.id", "demo-consumer-client");
        return prop;
    }

    private static Properties loadProps2(){
        final Properties prop = new Properties();
        prop.put("bootstrap.servers", "120.55.125.58:9092,120.26.198.248:9092,121.40.200.37:9092");
        prop.put("key.deserializer", "org.apache.kafka.common.serialization.StringDeserializer");
        prop.put("value.deserializer", "org.apache.kafka.common.serialization.StringDeserializer");
        prop.put("group.id","test_group4");
        prop.put("client.id", "demo-consumer-client");
        prop.put("auto.offset.reset", "earliest");
        prop.put("enable.auto.commit", "true"); //默认为true自动提交
        prop.put("auto.commit.interval.ms", "10000"); //设置默认自动提交时间 ,默认值为5000ms
        return prop;
    }
}
package com.hanwan.kafka.demo2;

import org.apache.kafka.clients.consumer.ConsumerRecords;
import org.apache.kafka.clients.consumer.KafkaConsumer;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

import java.util.Collections;
import java.util.Properties;
import java.util.concurrent.atomic.AtomicInteger;

/**
 * @ClassName ConsumerSyncCommit
 * @Description Consumer 同步commit
 * @Copyright: Copyright (c) 2018</p>
 * @Company: www.lowan.com</ p>
 * @Author hanwan
 * @Date 2018/8/20 10:07
 * @Version 1.0
 **/
public class ConsumerSyncCommit {

    private final static Logger LOGGER = LoggerFactory.getLogger(ConsumerSyncCommit.class);

    public static void main(String[] args) {
        KafkaConsumer<String, String> consumer = new KafkaConsumer<>(loadProps());
        consumer.subscribe(Collections.singleton("test12"));
        final AtomicInteger count = new AtomicInteger(0);
        while (true) {
            ConsumerRecords<String, String> records = consumer.poll(100);
            records.forEach(record ->{
                LOGGER.info("key {}", record.key());
                LOGGER.info("partition {}", record.partition());
                LOGGER.info("offset {}", record.offset());
                LOGGER.info("value {}", record.value());
//                if (count.incrementAndGet() == 100) {
//                    consumer.commitSync();
//                    count.set(0);
//                }
            });
            /*
            * can retry
            * block
            * */
            consumer.commitSync();
        }
    }

    private static Properties loadProps(){
        final Properties prop = new Properties();
        prop.put("bootstrap.servers", "120.55.125.58:9092,120.26.198.248:9092,121.40.200.37:9092");
        prop.put("key.deserializer", "org.apache.kafka.common.serialization.StringDeserializer");
        prop.put("value.deserializer", "org.apache.kafka.common.serialization.StringDeserializer");
        prop.put("group.id", "test_group1");
        prop.put("client.id", "demo-consumer-client");
        prop.put("auto.offset.reset", "earliest");
        prop.put("enable.auto.commit", "false");
        return prop;
    }
}
package com.hanwan.kafka.demo2;

import org.apache.kafka.clients.consumer.ConsumerRecords;
import org.apache.kafka.clients.consumer.KafkaConsumer;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

import java.util.Collections;
import java.util.Properties;

/**
 * @ClassName ConsumerAsyncCommit
 * @Description Consumer 异步commit
 * @Copyright: Copyright (c) 2018</p>
 * @Company: www.lowan.com</ p>
 * @Author hanwan
 * @Date 2018/8/20 13:28
 * @Version 1.0
 **/
public class ConsumerAsyncCommit {

    private final static Logger LOGGER = LoggerFactory.getLogger(ConsumerAsyncCommit.class);

    public static void main(String[] args) {
        KafkaConsumer<String, String> consumer = new KafkaConsumer<>(loadProp());
        consumer.subscribe(Collections.singleton("test12"));
        while (true) {
            ConsumerRecords<String, String> records = consumer.poll(100);
            records.forEach(record ->{
                LOGGER.info("key {}", record.key());
                LOGGER.info("partition {}", record.partition());
                LOGGER.info("offset {}", record.offset());
                LOGGER.info("value {}", record.value());
          
  • 0
    点赞
  • 2
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值