flink table api/sql消费kafka的json数据保存到mysql

flink消费kafka数据的版本问题,可以去https://mvnrepository.com/,查看对应版本。
环境:
mysql
zookeeper:3.4.13
kafka:0.8_2.11
flink:1.7.2(pom.xml中)

完整代码:
pom.xml:
在这里插入图片描述
在这里插入图片描述
在这里插入图片描述

代码FlinkKafkajson:

import org.apache.flink.api.common.typeinfo.Types;
import org.apache.flink.streaming.api.datastream.DataStream;
import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
import org.apache.flink.table.api.Table;
import org.apache.flink.table.api.TableEnvironment;
import org.apache.flink.table.api.java.StreamTableEnvironment;
import org.apache.flink.table.descriptors.Json;
import org.apache.flink.table.descriptors.*;
import org.apache.flink.table.descriptors.Schema;
import org.apache.flink.types.Row;

/**
 * Created by lhq on 2019/6/11.
 */
public  class FlinkKafkajson {
    public static void main(String[] args) throws Exception {
        StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();
        StreamTableEnvironment tableEnvironment = TableEnvironment.getTableEnvironment(env);
        tableEnvironment.connect(new Kafka().version("0.8").topic("lhqtest").startFromLatest()
                .property("bootstrap.servers","192.168.x.x:9092")
                .property("zookeeper.connect","192.168.x.x:2181")
                .property("group.id", "lhqtest"))
                .withFormat(new Json().failOnMissingField(true).deriveSchema())
                .withSchema(new Schema()
                        .field("id", Types.INT)
                        .field("name", Types.STRING)
                        .field("sex", Types.STRING)

                )
                .inAppendMode()
                .registerTableSource("lhq_user");
        Table table = tableEnvironment.scan("lhq_user").select("id,name,sex");
        DataStream<Row> personDataStream = tableEnvironment.toAppendStream(table,Row.class);
        personDataStream.addSink(new MysqlSink());
        env.execute("userPv from Kafka");

    }
}

写入mysql代码:

import org.apache.flink.api.java.tuple.Tuple2;
import org.apache.flink.configuration.Configuration;
import org.apache.flink.streaming.api.datastream.DataStream;
import org.apache.flink.streaming.api.functions.sink.RichSinkFunction;
import org.apache.flink.types.Row;

import java.sql.Connection;
import java.sql.DriverManager;
import java.sql.PreparedStatement;

/**
 * Created by luhaiqing on 2019/6/5.
 */

    public class MysqlSink extends RichSinkFunction<Row>
    {

        private Connection connection;
        private PreparedStatement preparedStatement;

        @Override
        public void open(Configuration parameters) throws Exception {
            super.open(parameters);
            String className = "com.mysql.jdbc.Driver";
            Class.forName(className);
            String url = "jdbc:mysql://localhost:3306/test";
            String user = "root";
            String password = "123456";
            connection = DriverManager.getConnection(url, user, password);
            String sql = "replace into flinkjsontest(id,name,sex) values(?,?,?)";
            preparedStatement = connection.prepareStatement(sql);
            super.open(parameters);
        }

        @Override
        public void close() throws Exception {
            super.close();
            if (preparedStatement != null) {
                preparedStatement.close();
            }
            if (connection != null) {
                connection.close();
            }
            super.close();
        }

        public void invoke(Row value, Context context) throws Exception {

            int   id = (int)value.getField(0);
            String   name = (String)value.getField(1);
            String   sex = (String)value.getField(2);

            System.out.print(id+":"+name+":"+sex);
            preparedStatement.setInt(1, id);
            preparedStatement.setString(2, name);
            preparedStatement.setString(3,sex);
            int i = preparedStatement.executeUpdate();
            if (i > 0) {
                System.out.println("value=" + value);
            }
        }

    }
  • 3
    点赞
  • 12
    收藏
    觉得还不错? 一键收藏
  • 5
    评论
1. 环境准备 - 安装 MySQL,创建测试数据库和表,并插入数据 - 安装 Kafka,并创建一个 topic - 安装 Flink 2. 创建 Flink 项目 - 在 Flink 的 bin 目录下执行 flink new myflinkproject 创建一个新的 Flink 项目 - 在 pom.xml 中添加以下依赖 ``` <dependency> <groupId>org.apache.flink</groupId> <artifactId>flink-table-api-java-bridge</artifactId> <version>${flink.version}</version> </dependency> <dependency> <groupId>org.apache.flink</groupId> <artifactId>flink-connector-jdbc</artifactId> <version>${flink.version}</version> </dependency> <dependency> <groupId>org.apache.flink</groupId> <artifactId>flink-connector-kafka_2.11</artifactId> <version>${flink.version}</version> </dependency> <dependency> <groupId>org.apache.flink</groupId> <artifactId>flink-json</artifactId> <version>${flink.version}</version> </dependency> ``` - 在 src/main/java 下创建一个 Java 类,例如 SyncMySQLToKafka.java 3. 编写 Flink SQL 在 SyncMySQLToKafka.java 中编写以下代码: ``` public class SyncMySQLToKafka { public static void main(String[] args) throws Exception { StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment(); EnvironmentSettings settings = EnvironmentSettings.newInstance().useBlinkPlanner().inStreamingMode().build(); StreamTableEnvironment tableEnv = StreamTableEnvironment.create(env, settings); tableEnv.executeSql("CREATE TABLE mysql_table (id INT, name STRING) " + "WITH (" + " 'connector.type' = 'jdbc'," + " 'connector.url' = 'jdbc:mysql://localhost:3306/test?characterEncoding=utf-8'," + " 'connector.table' = 'test_table'," + " 'connector.driver' = 'com.mysql.jdbc.Driver'," + " 'connector.username' = 'root'," + " 'connector.password' = 'root'" + ")"); tableEnv.executeSql("CREATE TABLE kafka_table (id INT, name STRING) " + "WITH (" + " 'connector.type' = 'kafka'," + " 'connector.version' = 'universal'," + " 'connector.topic' = 'test_topic'," + " 'connector.properties.bootstrap.servers' = 'localhost:9092'," + " 'connector.properties.group.id' = 'test_group'," + " 'format.type' = 'json'," + " 'update-mode' = 'append'" + ")"); tableEnv.executeSql("INSERT INTO kafka_table SELECT id, name FROM mysql_table"); env.execute(); } } ``` - 创建一个 MySQLmysql_table,指定连接信息和表名 - 创建一个 Kafkakafka_table,指定连接信息、topic 和数据格式 - 将 mysql_table 中的数据插入到 kafka_table 中 4. 运行程序 - 在命令行中进入项目根目录,执行 mvn clean package 编译项目 - 执行以下命令运行程序 ``` ./bin/flink run -c SyncMySQLToKafka target/myflinkproject-1.0-SNAPSHOT.jar ``` 5. 验证结果 - 在 Kafka 中查看是否有数据写入到 test_topic 中 - 修改 MySQL 表中的数据,查看是否能同步到 Kafka 中 以上就是使用 Flink SQL 实现 MySQL 同步到 Kafka 的简单示例。需要注意的是,本示例仅供参考,实际应用中需要根据具体需求进行修改和优化。
评论 5
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值