大数据学习hadoop3.1.3——Flume企业开发案例三(聚合)

1、案例需求:

hadoop101上的Flume-1监控文件/opt/module/group.log,
hadoop102上的Flume-2监控某一个端口的数据流,
Flume-1与Flume-2将数据发送给hadoop103上的Flume-3,Flume-3将最终数据打印到控制台。

2、需求分析
在这里插入图片描述
3、实现步骤:

(1)准备工作

分发Flume

 xsync flume

在hadoop101、hadoop102以及hadoop103的/opt/module/flume/job目录下创建一个group3文件夹。

 mkdir group3

(2)创建flume1-logger-flume.conf

配置Source用于监控hive.log文件,配置Sink输出数据到下一级Flume。
在hadoop101上编辑配置文件

 vim flume1-logger-flume.conf 

添加如下内容

 # Name the components on this agent
 a1.sources = r1
 a1.sinks = k1
 a1.channels = c1
 
 # Describe/configure the source
 a1.sources.r1.type = exec
 a1.sources.r1.command = tail -F /opt/module/group.log
 a1.sources.r1.shell = /bin/bash -c
 
 # Describe the sink
 a1.sinks.k1.type = avro
 a1.sinks.k1.hostname = hadoop103
 a1.sinks.k1.port = 4141
 
 #Describe the channel
 a1.channels.c1.type = memory
 a1.channels.c1.capacity = 1000
 a1.channels.c1.transactionCapacity = 100
 
 # Bind the source and sink to the channel
 a1.sources.r1.channels = c1
 a1.sinks.k1.channel = c1

(3)创建flume2-netcat-flume.conf

配置Source监控端口44444数据流,配置Sink数据到下一级Flume:

在hadoop102上编辑配置文件

 vim flume2-netcat-flume.conf

添加如下内容

 # Name the components on this agent
 a2.sources = r1
 a2.sinks = k1
 a2.channels = c1
 
 # Describe/configure the source
 a2.sources.r1.type = netcat
 a2.sources.r1.bind = hadoop102
 a2.sources.r1.port = 44444
 
 # Describe the sink
 a2.sinks.k1.type = avro
 a2.sinks.k1.hostname = hadoop103
 a2.sinks.k1.port = 4141
 
 # Use a channel which buffers events in memory
 a2.channels.c1.type = memory
 a2.channels.c1.capacity = 1000
 a2.channels.c1.transactionCapacity = 100
 
 # Bind the source and sink to the channel
 a2.sources.r1.channels = c1
 a2.sinks.k1.channel = c1

(4)创建flume3-flume-logger.conf

配置source用于接收flume1与flume2发送过来的数据流,最终合并后sink到控制台。

在hadoop103上编辑配置文件

 touch flume3-flume-logger.conf
 
 vim flume3-flume-logger.conf

添加如下内容

 # Name the components on this agent
 a3.sources = r1
 a3.sinks = k1
 a3.channels = c1
 
 # Describe/configure the source
 a3.sources.r1.type = avro
 a3.sources.r1.bind = hadoop103
 a3.sources.r1.port = 4141
 
 # Describe the sink
 # Describe the sink
 a3.sinks.k1.type = logger
 
 # Describe the channel
 a3.channels.c1.type = memory
 a3.channels.c1.capacity = 1000
 a3.channels.c1.transactionCapacity = 100
 
 # Bind the source and sink to the channel
 a3.sources.r1.channels = c1
 a3.sinks.k1.channel = c1

(5)执行配置文件

分别开启对应配置文件:flume3-flume-logger.conf,flume2-netcat-flume.conf,flume1-logger-flume.conf。
103:

 bin/flume-ng agent --conf conf/ --name a3 --conf-file job/group3/flume3-flume-logger.conf -Dflume.root.logger=INFO,console

101:

 bin/flume-ng agent --conf conf/ --name a1 --conf-file job/group3/flume1-logger-flume.conf

102:

 bin/flume-ng agent --conf conf/ --name a2 --conf-file job/group3/flume2-netcat-flume.conf

(6)在hadoop101上向/opt/module目录下的group.log追加内容

 echo 'hello' > group.log

(7)在hadoop102上向44444端口发送数据

 nc hadoop103 44444

(8)检查hadoop103上数据

评论 2
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值