Flume Source官网剖析(博主推荐)

 

 

  

   不多说,直接上干货!

 

 

 

 

 

  一切来源于flume官网

http://flume.apache.org/FlumeUserGuide.html

 

 

 

 

 

Flume Sources

 

   

 

 

 

Avro Source

 

 

   官网给的例子是

a1.sources = r1
a1.channels = c1
a1.sources.r1.type = avro
a1.sources.r1.channels = c1
a1.sources.r1.bind = 0.0.0.0
a1.sources.r1.port = 4141

 

 

 

 

 而我们常用的一般是,

agent1.sources = avro-source1
agent1.channels = ch1



#Define and configure an Spool directory source
agent1.sources.avro-source1.channels=ch1
agent1.sources.avro-source1.type=avro
agent1.sources.avro-source1.bind=0.0.0.0 agent1.sources.avro-source1.port=4141


a1.sources.r1.interceptors = i1
a1.sources.r1.interceptors.i1.type = com.djt.flume.interceptor.BehaviorIterceptor$BehaviorIterceptorBuilder


 

 

 

 

 

 

 

   

Thrift Source

 

 

 

 

 

 

  

Exec Source

 

 

 

 

 

 

 

 

 

JMS Source

 

 

 

 

 

 

 

 

 

 

Spooling Directory Source(常用)

 

 

  官网上给的参考例子是

a1.channels = ch-1
a1.sources = src-1

a1.sources.src-1.type = spooldir
a1.sources.src-1.channels = ch-1
a1.sources.src-1.spoolDir = /var/log/apache/flumeSpool
a1.sources.src-1.fileHeader = true

 

 

  而我们常用的一般是,

agent1.sources = spool-source1
agent1.channels = ch1



#Define and configure an Spool directory source
agent1.sources.spool-source1.channels=ch1
agent1.sources.spool-source1.type=spooldir
agent1.sources.spool-source1.spoolDir=/home/hadoop/data/flume/sqooldir
agent1.sources.spool-source1.ignorePattern=event(_\d{4}\-d{2}\-d{2}\_d{2}\_d{2})?\.log(\.COMPLETED)?
agent1.sources.spool-source1.deserializer.maxLineLength=10240

 

 

 

 

 

 

 

Taildir Source

 

 

 

 

 

 

 

 

Twitter 1% firehose Source

 

 

 

 

 

 

 

Kafka Source(常用)

 

   官网给的例子是

Example for topic subscription by comma-separated topic list.

tier1.sources.source1.type = org.apache.flume.source.kafka.KafkaSource
tier1.sources.source1.channels = channel1
tier1.sources.source1.batchSize = 5000
tier1.sources.source1.batchDurationMillis = 2000
tier1.sources.source1.kafka.bootstrap.servers = localhost:9092
tier1.sources.source1.kafka.topics = test1, test2
tier1.sources.source1.kafka.consumer.group.id = custom.g.id

 

 

 Example for topic subscription by regex

tier1.sources.source1.type = org.apache.flume.source.kafka.KafkaSource
tier1.sources.source1.channels = channel1
tier1.sources.source1.kafka.bootstrap.servers = localhost:9092
tier1.sources.source1.kafka.topics.regex = ^topic[0-9]$
# the default kafka.consumer.group.id=flume is used

 

  具体官网里,还给了Security and Kafka Source、TLS and Kafka Source、Kerberos and Kafka Source。自行去看吧

 

 

 

 

 

 

 

 

NetCat Source

 

 

 

 

 

 

 

 

 

Sequence Generator Source

 

 

 

 

 

 

Syslog Sources

 

 

 

 

 

 

 

 

 

Syslog TCP Source

 

 

 

 

 

 

 

 

 

 

 

 

 

 

Multiport Syslog TCP Source

 

 

 

 

 

 

 

 

 

 

Syslog UDP Source

 

 

 

 

 

 

 

 

HTTP Source

 

   官网还提供了JSONHandler、BlobHandler。这里不多说,自行去看吧

 

 

 

 

 

 

Stress Source

 

 

 

 

 

 

 

Legacy Sources

 

 

 

 

 

 

 

 

 

 

 

Avro Legacy Source

 

 

 

 

 

 

 

 

 

 

 

Thrift Legacy Source

 

 

 

 

 

 

 

 

 

 

 

 

Custom Source

 

 

 

 

 

 

 

 

Scribe Source

 

 

   

 大家自己去看,罗列出来,是让大家,不要局限于博客本身,眼光要放宽和多看官网,一切来源于官网。

  对于大数据无论是各种开源项目,看官网是学习的最好办法,再加上他人的中文博客。不要觉得英文恐惧,专业英语也就那样!变成高手和大牛,加油吧,zhouls!

 

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值