1 maven依赖
<dependency>
<groupId>org.springframework.kafka</groupId>
<artifactId>spring-kafka</artifactId>
<version>2.1.0.RELEASE</version>
</dependency>
<dependency>
<groupId>org.apache.flink</groupId>
<artifactId>flink-connector-kafka_2.11</artifactId>
<version>1.10.0</version>
</dependency>
<dependency>
<groupId>org.apache.flink</groupId>
<artifactId>flink-java</artifactId>
<version>1.10.0</version>
</dependency>
<dependency>
<groupId>org.apache.flink</groupId>
<artifactId>flink-streaming-java_2.11</artifactId>
<version>1.10.0</version>
</dependency>
<dependency>
<groupId>org.apache.flink</groupId>
<artifactId>flink-clients_2.11</artifactId>
<version>1.10.0</version>
</dependency>
<dependency>
<groupId>joda-time</groupId>
<artifactId>joda-time</artifactId>
<version>2.8.1</version>
</dependency>
<dependency>
<groupId>com.alibaba</groupId>
<artifactId>fastjson</artifactId>
<version>1.2.73</version>
</dependency>
<dependency>
<groupId>org.projectlombok</groupId>
<artifactId>lombok</artifactId>
<optional>true</optional>
</dependency>
<dependency>
<groupId>mysql</groupId>
<artifactId>mysql-connector-java</artifactId>
<version>5.1.34</version>
</dependency>
<dependency>
<groupId>org.apache.commons</groupId>
<artifactId>commons-dbcp2</artifactId>
<version>2.1.1</version>
</dependency>
2 消息生产者
import com.alibaba.fastjson.JSON;
import com.test.domain.MessageEntity;
import lombok.extern.slf4j.Slf4j;
import org.joda.time.DateTime;
import org.springframework.stereotype.Component;
import org.apache.kafka.clients.producer.KafkaProducer;
import org.apache.kafka.clients.producer.ProducerRecord;
import org.apache.kafka.clients.producer.RecordMetadata;
import java.util.Properties;
import java.util.concurrent.ExecutionException;
import java.util.concurrent.Future;
@Slf4j
@Component
public class MessageProducter {
private static Properties getProps(){
Properties props = new Properties();
props.put("bootstrap.servers", "192.168.9.10:9092");
props.put("acks", "all"); // 发送所有ISR
props.put("retries", Integer.MAX_VALUE); // 重试次数
props.put("batch.size", 16384); // 批量发送大小
props.put("buffer.memory", 102400); // 缓存大小,根据本机内存大小配置
props.put("linger.ms", 1000); // 发送频率,满足任务一个条件发送
props.put("client.id", "producer-syn-1"); // 发送端id,便于统计
props.put("key.serializer", "org.apache.kafka.common.serialization.StringSerializer");
props.put("value.serializer", "org.apache.kafka.common.serialization.StringSerializer");
return props;
}
public static void main(String[] args) throws Exception {
KafkaProducer<String, String> producer = new KafkaProducer<>(getProps());
for (int i=30;i<=100;i++){
MessageEntity message = new MessageEntity();
message.setMessage("第"+i+"条:message");
message.setTotalDate(DateTime.now().toString("yyyy-MM-dd HH:mm:ss"));
message.setMessageId(System.currentTimeMillis());
log.info(JSON.toJSONString(message));
ProducerRecord<String, String> record = new ProducerRecord<>("flink_test_01",
System.currentTimeMillis()+"", JSON.toJSONString(message));
Future<RecordMetadata> metadataFuture = producer.send(record);
RecordMetadata recordMetadata;
try {
recordMetadata = metadataFuture.get();
log.info("发送成功!");
log.info("topic:"+recordMetadata.topic());
log.info("partition:"+recordMetadata.partition());
log.info("offset:"+recordMetadata.offset());
} catch (InterruptedException|ExecutionException e) {
System.out.println("发送失败!");
e.printStackTrace();
}
Thread.sleep(10000);
}
}
}
./kafka-console-consumer.sh --bootstrap-server localhost:9092 --topic flink_test_01 --from-beginning
3 从kafka中读取数据,转成Message实体来写入
public class KafkaToMysql {
public static void main(String[] args) throws Exception{
// 构建环境
StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();
//指定checkpoint的触发间隔
env.enableCheckpointing(5000);
env.setParallelism(1);
env.getCheckpointConfig().setCheckpointingMode(CheckpointingMode.EXACTLY_ONCE);
Properties properties = new Properties();
properties.setProperty("bootstrap.servers", "192.168.0.96:9092");
properties.setProperty("group.id", "flink_consumer_01");
properties.setProperty("auto.offset.reset", "earliest");
SingleOutputStreamOperator<MessageEntity> streamOperator = env.addSource(
new FlinkKafkaConsumer<>("flink_test_01",
new MessageDeSerializationSchema(), properties)).setParallelism(1)
.map(record -> {
MessageEntity message = null;
try {
System.out.println("consumerRecord"+record.value());
message = JSON.parseObject(record.value(), MessageEntity.class);
System.out.println(message.toString());
}catch (Exception e){
System.out.println("格式问题:"+ record);
}
return message;
});
SingleOutputStreamOperator<List<MessageEntity>> process = streamOperator.timeWindowAll(Time.seconds(10))
.process(new AllWindownFunction1());
process.addSink(new SinkToMysql()).setParallelism(1).name("flink_test_01");
env.execute("consumer start");
}
public static class AllWindownFunction1 extends ProcessAllWindowFunction<MessageEntity, List<MessageEntity>, TimeWindow> {
@Override
public void process(ProcessAllWindowFunction<MessageEntity, List<MessageEntity>,
TimeWindow>.Context context, Iterable<MessageEntity> iterable,
Collector<List<MessageEntity>> collector) throws Exception {
if (iterable != null){
System.out.println("非空集合"+iterable);
ArrayList<MessageEntity> list = Lists.newArrayList(iterable);
if (list.size() > 0) {
System.out.println("10s数据条数:" + list.size());
collector.collect(list);
}
}else {
System.out.println("空集合");
}
}
}
}
@Data
public class MessageEntity {
private Long messageId;
private String message;
private String totalDate;
}
4 mysqlSink类
import com.test.domain.MessageEntity;
import org.apache.commons.dbcp2.BasicDataSource;
import org.apache.flink.configuration.Configuration;
import org.apache.flink.streaming.api.functions.sink.RichSinkFunction;
import java.sql.Connection;
import java.sql.PreparedStatement;
import java.util.List;
public class SinkToMysql extends RichSinkFunction<List<MessageEntity>> {
PreparedStatement ps;
BasicDataSource dataSource;
private Connection connection;
@Override
public void open(Configuration parameters) throws Exception {
System.out.println("1111111111111");
dataSource = new BasicDataSource();
connection = getConnection(dataSource);
String sql = "insert into flink_test_01(message_id, message, total_date) values(?, ?, ?);";
ps = this.connection.prepareStatement(sql);
}
@Override
public void close() throws Exception {
super.close();
//关闭连接和释放资源
if (connection != null) {
connection.close();
}
if (ps != null) {
ps.close();
}
}
@Override
public void invoke(List<MessageEntity> value, Context context) throws Exception {
//遍历数据集合
for (MessageEntity message : value) {
ps.setLong(1, message.getMessageId());
ps.setString(2, message.getMessage());
System.out.println("message--------------------------:"+message.getMessage());
ps.setString(3, message.getTotalDate());
ps.addBatch();
}
int[] count = ps.executeBatch();//批量后执行
System.out.println("成功了插入了" + count.length + "行数据");
}
private static Connection getConnection(BasicDataSource dataSource) {
dataSource.setDriverClassName("com.mysql.jdbc.Driver");
dataSource.setUrl("jdbc:mysql://192.168.0.216:3306/hela?useUnicode=true&characterEncoding=utf8");
dataSource.setUsername("root");
dataSource.setPassword("123456");
dataSource.setInitialSize(10);
dataSource.setMaxTotal(50);
dataSource.setMinIdle(2);
Connection con = null;
try {
con = dataSource.getConnection();
} catch (Exception e) {
System.out.println("msg:" + e.getMessage());
}
return con;
}
}
运行效果: