flume 190 INSTALL

#1、解压和改名

tar  -zxvf /opt/download/hadoop/apache-flume-1.9.0-bin.tar.gz -C /opt/software/
mv /opt/software/apache-flume-1.9.0-bin /opt/software/flume190

#2、配置

cd /opt/software/flume190/conf
mv flume-env.sh.template flume-env.sh
export JAVA_HOME=/opt/software/java/jdk180

#配置环境变量并激活
vim /etc/profile.d/myenv.sh
#flume
# flume
export FLUME_HOME=/opt/software/flume190
export FLUME_CONF_DIR=$FLUME_HOME/conf
export FLUME_LOG_DIR=$FLUME_HOME/log
export PATH=$PATH:$FLUME_HOME/bin

source /etc/profile

##google的guava-xxx.jar和hadoop的版本是否一致

在flume190/lib下删除原来的guava
cp	/opt/software/hadoop/hadoop313/share/hadoop/hdfs/lib/guava-27.0-jre.jar在这里插入代码片

#flume

Event
SOURCE 				=> 			CHANNEL 			=> 			SINK
NETCAT(iP+port)					memory							logger
SPOOLDIR(dir)					file							hdfs
TAILDIR(dir)													kafka
	
yum -y install gcc
yum install -y nc

#启动flume

flume-ng agent --name a1 --conf /opt/software/flume190/conf/ --conf-file /root/flume_job/logconf/flume01.cnf -Dflume.root.logger=INFO,console
	
[root@singlelucky ~]# nc -v 192.168.19.130 6666
Ncat: Version 7.50 ( https://nmap.org/ncat )
Ncat: Connected to 192.168.19.130:6666.
SOURCE:NETCAT,CHANNEL:MEMORT,SINK:LOGGER

#vim flume01.cnf

#组件声明
a1.sources = s1
a1.channels = c1
a1.sinks = k1
	
#初始化数据源
a1.sources.s1.type = netcat
a1.sources.s1.bind = 192.168.19.130
a1.sources.s1.port = 6666
	
#初始化通道
a1.channels.c1.type = memory
a1.channels.c1.capacity = 100
a1.channels.c1.transacionCapacity = 10
	
#初始化数据槽
a1.sinks.k1.type = logger
	
#关联组件
a1.sources.s1.channels = c1
a1.sinks.k1.channel = c1
  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值