flink 四 kafka sink mysql

621 阅读2分钟

1 maven依赖

        <dependency>
            <groupId>org.springframework.kafka</groupId>
            <artifactId>spring-kafka</artifactId>
            <version>2.1.0.RELEASE</version>
        </dependency>

        <dependency>
            <groupId>org.apache.flink</groupId>
            <artifactId>flink-connector-kafka_2.11</artifactId>
            <version>1.10.0</version>
        </dependency>

        <dependency>
            <groupId>org.apache.flink</groupId>
            <artifactId>flink-java</artifactId>
            <version>1.10.0</version>
        </dependency>
        <dependency>
            <groupId>org.apache.flink</groupId>
            <artifactId>flink-streaming-java_2.11</artifactId>
            <version>1.10.0</version>
        </dependency>
        <dependency>
            <groupId>org.apache.flink</groupId>
            <artifactId>flink-clients_2.11</artifactId>
            <version>1.10.0</version>
        </dependency>
        <dependency>
            <groupId>joda-time</groupId>
            <artifactId>joda-time</artifactId>
            <version>2.8.1</version>
        </dependency>
        <dependency>
            <groupId>com.alibaba</groupId>
            <artifactId>fastjson</artifactId>
            <version>1.2.73</version>
        </dependency>
        <dependency>
            <groupId>org.projectlombok</groupId>
            <artifactId>lombok</artifactId>
            <optional>true</optional>
        </dependency>
        <dependency>
            <groupId>mysql</groupId>
            <artifactId>mysql-connector-java</artifactId>
            <version>5.1.34</version>
        </dependency>

        <dependency>
            <groupId>org.apache.commons</groupId>
            <artifactId>commons-dbcp2</artifactId>
            <version>2.1.1</version>
        </dependency>

2 消息生产者

import com.alibaba.fastjson.JSON;
import com.quwan.domain.MessageEntity;
import lombok.extern.slf4j.Slf4j;
import org.joda.time.DateTime;
import org.springframework.stereotype.Component;
import org.apache.kafka.clients.producer.KafkaProducer;
import org.apache.kafka.clients.producer.ProducerRecord;
import org.apache.kafka.clients.producer.RecordMetadata;

import java.util.Properties;
import java.util.concurrent.ExecutionException;
import java.util.concurrent.Future;


@Slf4j
@Component
public class MessageProducter {

    private static Properties getProps(){
        Properties props =  new Properties();
        props.put("bootstrap.servers", "10.112.192.96:9092");
        props.put("acks", "all"); // 发送所有ISR
        props.put("retries", Integer.MAX_VALUE); // 重试次数
        props.put("batch.size", 16384); // 批量发送大小
        props.put("buffer.memory", 102400); // 缓存大小,根据本机内存大小配置
        props.put("linger.ms", 1000); // 发送频率,满足任务一个条件发送
        props.put("client.id", "producer-syn-1"); // 发送端id,便于统计
        props.put("key.serializer", "org.apache.kafka.common.serialization.StringSerializer");
        props.put("value.serializer", "org.apache.kafka.common.serialization.StringSerializer");
        return props;
    }


    public static void main(String[] args) throws Exception {
        KafkaProducer<String, String> producer = new KafkaProducer<>(getProps());
        for (int i=30;i<=100;i++){
            MessageEntity message = new MessageEntity();
            message.setMessage("第"+i+"条:message");
            message.setTotalDate(DateTime.now().toString("yyyy-MM-dd HH:mm:ss"));
            message.setMessageId(System.currentTimeMillis());
            log.info(JSON.toJSONString(message));
            ProducerRecord<String, String> record = new ProducerRecord<>("flink_test_01",
                    System.currentTimeMillis()+"", JSON.toJSONString(message));
            Future<RecordMetadata> metadataFuture = producer.send(record);
            RecordMetadata recordMetadata;
            try {
                recordMetadata = metadataFuture.get();
                log.info("发送成功!");
                log.info("topic:"+recordMetadata.topic());
                log.info("partition:"+recordMetadata.partition());
                log.info("offset:"+recordMetadata.offset());
            } catch (InterruptedException|ExecutionException e) {
                System.out.println("发送失败!");
                e.printStackTrace();
            }
            Thread.sleep(10000);
        }
    }
}


./kafka-console-consumer.sh --bootstrap-server localhost:9092 --topic flink_test_01 --from-beginning 

3 从kafka中读取数据,转成Message实体来写入

public class KafkaToMysql {
    public static void main(String[] args) throws Exception{
        // 构建环境
        StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();
        //指定checkpoint的触发间隔
        env.enableCheckpointing(5000);
        env.setParallelism(1);
        env.getCheckpointConfig().setCheckpointingMode(CheckpointingMode.EXACTLY_ONCE);
        Properties properties = new Properties();
        properties.setProperty("bootstrap.servers", "10.112.192.96:9092");
        properties.setProperty("group.id", "flink_consumer_01");
        properties.setProperty("auto.offset.reset", "earliest");

        SingleOutputStreamOperator<MessageEntity> streamOperator = env.addSource(
                new FlinkKafkaConsumer<>("flink_test_01",
                        new MessageDeSerializationSchema(), properties)).setParallelism(1)
                .map(record -> {
                    MessageEntity message = null;
                    try {
                        System.out.println("consumerRecord"+record.value());
                        message = JSON.parseObject(record.value(), MessageEntity.class);
                        System.out.println(message.toString());
                    }catch (Exception e){
                        System.out.println("格式问题:"+ record);
                    }
                    return message;
                });

        SingleOutputStreamOperator<List<MessageEntity>> process = streamOperator.timeWindowAll(Time.seconds(10))
                .process(new AllWindownFunction1());
        process.addSink(new SinkToMysql()).setParallelism(1).name("flink_test_01");
        env.execute("consumer start");
    }

    public static class AllWindownFunction1 extends ProcessAllWindowFunction<MessageEntity, List<MessageEntity>, TimeWindow> {
        @Override
        public void process(ProcessAllWindowFunction<MessageEntity, List<MessageEntity>,
                TimeWindow>.Context context, Iterable<MessageEntity> iterable,
                            Collector<List<MessageEntity>> collector) throws Exception {
            if (iterable != null){
                System.out.println("非空集合"+iterable);
                ArrayList<MessageEntity> list = Lists.newArrayList(iterable);
                if (list.size() > 0) {
                    System.out.println("10s数据条数:" + list.size());
                    collector.collect(list);
                }
            }else {
                System.out.println("空集合");
            }
        }
    }
}



@Data
public class MessageEntity {
    private Long messageId;
    private String message;
    private String totalDate;
}

4 mysqlSink类

import com.quwan.domain.MessageEntity;
import org.apache.commons.dbcp2.BasicDataSource;
import org.apache.flink.configuration.Configuration;
import org.apache.flink.streaming.api.functions.sink.RichSinkFunction;

import java.sql.Connection;
import java.sql.PreparedStatement;
import java.util.List;

public class SinkToMysql extends RichSinkFunction<List<MessageEntity>> {

    PreparedStatement ps;
    BasicDataSource dataSource;
    private Connection connection;

    @Override
    public void open(Configuration parameters) throws Exception {
        System.out.println("1111111111111");
        dataSource = new BasicDataSource();
        connection = getConnection(dataSource);
        String sql = "insert into flink_test_01(message_id, message, total_date) values(?, ?, ?);";
        ps = this.connection.prepareStatement(sql);
    }

    @Override
    public void close() throws Exception {
        super.close();
        //关闭连接和释放资源
        if (connection != null) {
            connection.close();
        }
        if (ps != null) {
            ps.close();
        }
    }

    @Override
    public void invoke(List<MessageEntity> value, Context context) throws Exception {
        //遍历数据集合
        for (MessageEntity message : value) {
            ps.setLong(1, message.getMessageId());
            ps.setString(2, message.getMessage());
            System.out.println("message--------------------------:"+message.getMessage());
            ps.setString(3, message.getTotalDate());
            ps.addBatch();
        }
        int[] count = ps.executeBatch();//批量后执行
        System.out.println("成功了插入了" + count.length + "行数据");
    }

    private static Connection getConnection(BasicDataSource dataSource) {
        dataSource.setDriverClassName("com.mysql.jdbc.Driver");
        dataSource.setUrl("jdbc:mysql://10.112.30.216:3306/hela?useUnicode=true&characterEncoding=utf8");
        dataSource.setUsername("root");
        dataSource.setPassword("123456");
        dataSource.setInitialSize(10);
        dataSource.setMaxTotal(50);
        dataSource.setMinIdle(2);
        Connection con = null;
        
        try {
            con = dataSource.getConnection();
        } catch (Exception e) {
            System.out.println("msg:" + e.getMessage());
        }
        return con;
    }

}

运行效果: