#1、解压和改名
tar -zxvf /opt/download/hadoop/apache-flume-1.9.0-bin.tar.gz -C /opt/software/
mv /opt/software/apache-flume-1.9.0-bin /opt/software/flume190
#2、配置
cd /opt/software/flume190/conf
mv flume-env.sh.template flume-env.sh
export JAVA_HOME=/opt/software/java/jdk180
#配置环境变量并激活
vim /etc/profile.d/myenv.sh
#flume
# flume
export FLUME_HOME=/opt/software/flume190
export FLUME_CONF_DIR=$FLUME_HOME/conf
export FLUME_LOG_DIR=$FLUME_HOME/log
export PATH=$PATH:$FLUME_HOME/bin
source /etc/profile
##google的guava-xxx.jar和hadoop的版本是否一致
在flume190/lib下删除原来的guava
cp /opt/software/hadoop/hadoop313/share/hadoop/hdfs/lib/guava-27.0-jre.jar在这里插入代码片
#flume
Event
SOURCE => CHANNEL => SINK
NETCAT(iP+port) memory logger
SPOOLDIR(dir) file hdfs
TAILDIR(dir) kafka
yum -y install gcc
yum install -y nc
#启动flume
flume-ng agent --name a1 --conf /opt/software/flume190/conf/ --conf-file /root/flume_job/logconf/flume01.cnf -Dflume.root.logger=INFO,console
[root@singlelucky ~]# nc -v 192.168.19.130 6666
Ncat: Version 7.50 ( https://nmap.org/ncat )
Ncat: Connected to 192.168.19.130:6666.
SOURCE:NETCAT,CHANNEL:MEMORT,SINK:LOGGER
#vim flume01.cnf
#组件声明
a1.sources = s1
a1.channels = c1
a1.sinks = k1
#初始化数据源
a1.sources.s1.type = netcat
a1.sources.s1.bind = 192.168.19.130
a1.sources.s1.port = 6666
#初始化通道
a1.channels.c1.type = memory
a1.channels.c1.capacity = 100
a1.channels.c1.transacionCapacity = 10
#初始化数据槽
a1.sinks.k1.type = logger
#关联组件
a1.sources.s1.channels = c1
a1.sinks.k1.channel = c1