JavaApi 如何获取Kafka Topic数据量

废话不多说,直接上代码

import java.util.Collections;
import java.util.List;
import java.util.Map;
import java.util.Optional;
import java.util.Properties;
import java.util.stream.Collectors;
import org.apache.kafka.clients.consumer.KafkaConsumer;
import org.apache.kafka.common.TopicPartition;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

public class KafkaOffsetTools {
//    private final static Logger logger = LoggerFactory.getLogger(KafkaOffsetTools.class);

    public static long totalMessageCount(String brokerList,String topic) {
        Properties props = new Properties();
        props.put("bootstrap.servers", brokerList);
        props.put("group.id", "message_count_group");
        props.put("enable.auto.commit", "false");
        props.put("key.deserializer", "org.apache.kafka.common.serialization.StringDeserializer");
        props.put("value.deserializer", "org.apache.kafka.common.serialization.StringDeserializer");

        try (KafkaConsumer<String, String> consumer = new KafkaConsumer<>(props)) {
            List<TopicPartition> tps = Optional.ofNullable(consumer.partitionsFor(topic))
                    .orElse(Collections.emptyList())
                    .stream()
                    .map(info -> new TopicPartition(info.topic(), info.partition()))
                    .collect(Collectors.toList());
            Map<TopicPartition, Long> beginOffsets = consumer.beginningOffsets(tps);
            Map<TopicPartition, Long> endOffsets = consumer.endOffsets(tps);

            long sum = tps.stream().mapToLong(tp -> endOffsets.get(tp) - beginOffsets.get(tp)).sum();
            System.out.println("topic {"+topic+"} 的消息总量为: => " + sum);
            return sum;
        }

    }
  • 1
    点赞
  • 3
    收藏
    觉得还不错? 一键收藏
  • 1
    评论
要统计Kafka Topic每小时的数据量,您可以使用Kafka提供的Java API和时间相关的类来实现。以下是实现的步骤: 1. 创建KafkaConsumer对象并指定要订阅的topic。 ```java Properties props = new Properties(); props.put("bootstrap.servers", "localhost:9092"); props.put("group.id", "test-consumer-group"); props.put("enable.auto.commit", "true"); props.put("auto.commit.interval.ms", "1000"); props.put("key.deserializer", "org.apache.kafka.common.serialization.StringDeserializer"); props.put("value.deserializer", "org.apache.kafka.common.serialization.StringDeserializer"); KafkaConsumer<String, String> consumer = new KafkaConsumer<String, String>(props); consumer.subscribe(Arrays.asList("test-topic")); ``` 2. 创建一个HashMap来存储每小时的数据量。 ```java HashMap<String, Long> countMap = new HashMap<String, Long>(); ``` 3. 使用poll()方法从Kafka集群获取消息,并将每条消息的时间戳转换为小时级别的字符串。 ```java ConsumerRecords<String, String> records = consumer.poll(Duration.ofSeconds(10)); for (ConsumerRecord<String, String> record : records) { long timestamp = record.timestamp(); String hour = getHour(timestamp); if (countMap.containsKey(hour)) { long count = countMap.get(hour); countMap.put(hour, count + 1); } else { countMap.put(hour, 1L); } } ``` 其中getHour()方法可以通过时间戳获取小时级别的字符串: ```java private static String getHour(long timestamp) { SimpleDateFormat sdf = new SimpleDateFormat("yyyy-MM-dd-HH"); sdf.setTimeZone(TimeZone.getTimeZone("GMT+8")); Date date = new Date(timestamp); return sdf.format(date); } ``` 4. 输出每小时的数据量。 ```java for (Map.Entry<String, Long> entry : countMap.entrySet()) { System.out.println(entry.getKey() + ": " + entry.getValue()); } ``` 5. 关闭KafkaConsumer对象。 ```java consumer.close(); ``` 完整的示例代码如下: ```java import java.text.SimpleDateFormat; import java.time.Duration; import java.util.Arrays; import java.util.Date; import java.util.HashMap; import java.util.Map; import java.util.Properties; import java.util.TimeZone; import org.apache.kafka.clients.consumer.ConsumerRecord; import org.apache.kafka.clients.consumer.ConsumerRecords; import org.apache.kafka.clients.consumer.KafkaConsumer; public class KafkaTopicHourlyDataCount { public static void main(String[] args) { Properties props = new Properties(); props.put("bootstrap.servers", "localhost:9092"); props.put("group.id", "test-consumer-group"); props.put("enable.auto.commit", "true"); props.put("auto.commit.interval.ms", "1000"); props.put("key.deserializer", "org.apache.kafka.common.serialization.StringDeserializer"); props.put("value.deserializer", "org.apache.kafka.common.serialization.StringDeserializer"); KafkaConsumer<String, String> consumer = new KafkaConsumer<String, String>(props); consumer.subscribe(Arrays.asList("test-topic")); HashMap<String, Long> countMap = new HashMap<String, Long>(); ConsumerRecords<String, String> records = consumer.poll(Duration.ofSeconds(10)); for (ConsumerRecord<String, String> record : records) { long timestamp = record.timestamp(); String hour = getHour(timestamp); if (countMap.containsKey(hour)) { long count = countMap.get(hour); countMap.put(hour, count + 1); } else { countMap.put(hour, 1L); } } for (Map.Entry<String, Long> entry : countMap.entrySet()) { System.out.println(entry.getKey() + ": " + entry.getValue()); } consumer.close(); } private static String getHour(long timestamp) { SimpleDateFormat sdf = new SimpleDateFormat("yyyy-MM-dd-HH"); sdf.setTimeZone(TimeZone.getTimeZone("GMT+8")); Date date = new Date(timestamp); return sdf.format(date); } } ``` 注意:这段代码只能统计一小时内的数据量,如果要统计更长的时间段,需要将该代码放在一个无限循环中,并且在每一次循环中进行累加。
评论 1
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值