flume配置参考

1 redis source,memory channel,json intercepter and file_roll sink配置例子

agent.sources = redis redis1
agent.channels = memory
agent.sinks = fileSink


agent.sources.redis.type = com.flumeng.plugins.source.RedisLPopSource
agent.sources.redis.host = 192.168.121.128
agent.sources.redis.port = 6379
agent.sources.redis.eventNamePrefix = event
agent.sources.redis.timeout = 3000
agent.sources.redis.charset = utf-8
agent.sources.redis.interceptors = i1
agent.sources.redis.interceptors.i1.type = com.flumeng.plugins.source.JSONInterceptor$Builder


agent.sources.redis1.type = com.flumeng.plugins.source.RedisLRangeSource
agent.sources.redis1.host = 192.168.121.128
agent.sources.redis1.port = 6379
agent.sources.redis1.eventNamePrefix = event
agent.sources.redis1.timeout = 3000
agent.sources.redis1.charset = utf-8
agent.sources.redis1.eventbatchSize = 2000
agent.sources.redis1.interceptors = i2
agent.sources.redis1.interceptors.i2.type = com.flumeng.plugins.source.JSONInterceptor$Builder


agent.channels.memory.type = memory
agent.channels.memory.capacity = 1000000
agent.channels.memory.transactionCapacity = 100000


agent.sinks.fileSink.type = file_roll
agent.sinks.fileSink.channel = memory
agent.sinks.fileSink.sink.directory = /mnt/hgfs/share/result/
agent.sinks.fileSink.sink.rollInterval = 0
agent.sinks.fileSink.batchSize=500


agent.sources.redis.channels = memory
agent.sources.redis1.channels = memory


2 redis source,memory channel,json intercepter and avro sink配置例子

agent.sources = redis redis1
agent.channels = memory
agent.sinks = fileSink


agent.sources.redis.type = com.flumeng.plugins.source.RedisLPopSource
agent.sources.redis.host = 192.168.121.128
agent.sources.redis.port = 6379
agent.sources.redis.eventNamePrefix = event
agent.sources.redis.timeout = 3000
agent.sources.redis.charset = utf-8
agent.sources.redis.interceptors = i1
agent.sources.redis.interceptors.i1.type = com.flumeng.plugins.source.JSONInterceptor$Builder


agent.sources.redis1.type = com.flumeng.plugins.source.RedisLRangeSource
agent.sources.redis1.host = 192.168.121.128
agent.sources.redis1.port = 6379
agent.sources.redis1.eventNamePrefix = event
agent.sources.redis1.timeout = 3000
agent.sources.redis1.charset = utf-8
agent.sources.redis1.eventbatchSize = 2000
agent.sources.redis1.interceptors = i2
agent.sources.redis1.interceptors.i2.type = com.flumeng.plugins.source.JSONInterceptor$Builder


agent.channels.memory.type = memory
agent.channels.memory.capacity = 1000000
agent.channels.memory.transactionCapacity = 100000


agent.sinks.fileSink.type = file_roll
agent.sinks.fileSink.channel = memory
agent.sinks.fileSink.sink.directory = /mnt/hgfs/share/result/
agent.sinks.fileSink.sink.rollInterval = 0
agent.sinks.fileSink.batchSize=500


agent.sources.redis.channels = memory
agent.sources.redis1.channels = memory


3 avro source--memery channel ---HDFS sink---interceptors---serializers

agent.sources = avroSource
agent.channels = memory
agent.sinks = hdfsSink


agent.sources.avroSource.type = avro
agent.sources.avroSource.bind = 0.0.0.0
agent.sources.avroSource.port = 4141


agent.sources.avroSource.interceptors = i1
agent.sources.avroSource.interceptors.i1.type=regex_extractor
agent.sources.avroSource.interceptors.i1.regex = (\\d\\d\\d\\d-\\d\\d-\\d\\d\\s\\d\\d:\\d\\d)
agent.sources.avroSource.interceptors.i1.serializers = s1
agent.sources.avroSource.interceptors.i1.serializers.s1.type = org.apache.flume.interceptor.RegexExtractorInterceptorMillisSerializer
agent.sources.avroSource.interceptors.i1.serializers.s1.name = timestamp
agent.sources.avroSource.interceptors.i1.serializers.s1.pattern = yyyy-MM-dd HH:mm


agent.channels.memory.type = memory
agent.channels.memory.capacity = 1000000
agent.channels.memory.transactionCapacity = 100000


agent.sinks.hdfsSink.type = hdfs
agent.sinks.hdfsSink.channel = memory
agent.sinks.hdfsSink.hdfs.path = hdfs://littleNameservice/test/outputAnalysis/y=%Y/m=%m/d=%d/h=%H
agent.sinks.hdfsSink.hdfs.rollInterval =0 
agent.sinks.hdfsSink.hdfs.rollSize =268435456
agent.sinks.hdfsSink.hdfs.rollCount =0
#agent.sinks.hdfsSink.hdfs.filePrefix =%Y_%m_%d_%H
agent.sinks.hdfsSink.hdfs.filePrefix =event
agent.sinks.hdfsSink.hdfs.fileSuffix =.txt
agent.sinks.hdfsSink.hdfs.inUseSuffix =.tmp
agent.sinks.hdfsSink.hdfs.batchSize = 10000
agent.sinks.hdfsSink.hdfs.idleTimeout = 600
agent.sinks.hdfsSink.hdfs.fileType = DataStream


agent.sources.avroSource.channels = memory


4 avro source---memery channel---hbase sink---serializer---interceptor

agent.sources = avroSource
agent.channels = memory
agent.sinks = hbaseSink


agent.sources.avroSource.type = avro
agent.sources.avroSource.bind = 0.0.0.0
agent.sources.avroSource.port = 4141


agent.sources.avroSource.interceptors = i1
agent.sources.avroSource.interceptors.i1.type=com.jj.flume.interceptor.json.EventHbaseInterceptor$Builder


agent.channels.memory.type = memory
agent.channels.memory.capacity = 1000000
agent.channels.memory.transactionCapacity = 100000


agent.sinks.hbaseSink.type = hbase
agent.sinks.hbaseSink.table = t_event
agent.sinks.hbaseSink.columnFamily = f_event
agent.sinks.hbaseSink.serializer = com.jj.flume.sink.hbase.RegexHbaseEventSerializerJJ
agent.sinks.hbaseSink.channel = memory


agent.sources.avroSource.channels = memory


  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
配置Flume与Kafka的集成,可以按照以下步骤进行操作: 1. 首先,需要在Flume配置文件中定义sourcesinkchannel。其中,source指定为Kafka的source端,sink指定为Kafka的sink端,channel用于在sourcesink之间传递数据。配置文件中的示例配置可以参考引用中的flume-kafka.conf。 2. 如果要将Kafka作为source端,需要在配置文件中设置source的类型为kafka,并指定Kafka的相关参数,如Kafka的地址、topic名称等。 3. 启动Flume之前,确保Zookeeper和Kafka已经成功启动。因为Flume在启动时会连接Kafka,如果Kafka未启动,会导致连接错误。参考引用中的说明。 4. 如果要将Kafka作为sink端,需要在配置文件中设置sink的类型为kafka,并指定Kafka的相关参数,如Kafka的地址、topic名称等。 5. 在启动Flume之前,需要确保Zookeeper集群和Kafka集群已经成功启动。 6. 可以使用命令行创建和查看Kafka的topic,以及创建Kafka的消费者。具体命令可以根据实际需求进行设置。 7. 启动Flume,并发送消息到Flume的端口。 8. 在Kafka的消费者中接收消息,验证数据是否成功传输。 需要注意的是,在配置Flume和Kafka的过程中,需要根据实际情况调整相关参数,确保Flume和Kafka能够正确地进行数据传输和接收。配置文件的具体内容可以参考引用中的示例配置。<span class="em">1</span><span class="em">2</span><span class="em">3</span> #### 引用[.reference_title] - *1* [Flume对接Kafka详细过程](https://blog.csdn.net/qq_47183158/article/details/112179052)[target="_blank" data-report-click={"spm":"1018.2226.3001.9630","extra":{"utm_source":"vip_chatgpt_common_search_pc_result","utm_medium":"distribute.pc_search_result.none-task-cask-2~all~insert_cask~default-1-null.142^v93^chatsearchT3_2"}}] [.reference_item style="max-width: 33.333333333333336%"] - *2* [玩转Flume+Kafka原来也就那点事儿](https://blog.csdn.net/weixin_34189116/article/details/89658220)[target="_blank" data-report-click={"spm":"1018.2226.3001.9630","extra":{"utm_source":"vip_chatgpt_common_search_pc_result","utm_medium":"distribute.pc_search_result.none-task-cask-2~all~insert_cask~default-1-null.142^v93^chatsearchT3_2"}}] [.reference_item style="max-width: 33.333333333333336%"] - *3* [快速学习-Flume 对接 Kafka](https://download.csdn.net/download/weixin_38592548/14885746)[target="_blank" data-report-click={"spm":"1018.2226.3001.9630","extra":{"utm_source":"vip_chatgpt_common_search_pc_result","utm_medium":"distribute.pc_search_result.none-task-cask-2~all~insert_cask~default-1-null.142^v93^chatsearchT3_2"}}] [.reference_item style="max-width: 33.333333333333336%"] [ .reference_list ]
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值