flink datasource
This commit is contained in:
		
							
								
								
									
										242
									
								
								code/Flink/flink-kafka-integration/pom.xml
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										242
									
								
								code/Flink/flink-kafka-integration/pom.xml
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,242 @@ | ||||
| <!-- | ||||
| Licensed to the Apache Software Foundation (ASF) under one | ||||
| or more contributor license agreements.  See the NOTICE file | ||||
| distributed with this work for additional information | ||||
| regarding copyright ownership.  The ASF licenses this file | ||||
| to you under the Apache License, Version 2.0 (the | ||||
| "License"); you may not use this file except in compliance | ||||
| with the License.  You may obtain a copy of the License at | ||||
|  | ||||
|   http://www.apache.org/licenses/LICENSE-2.0 | ||||
|  | ||||
| Unless required by applicable law or agreed to in writing, | ||||
| software distributed under the License is distributed on an | ||||
| "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY | ||||
| KIND, either express or implied.  See the License for the | ||||
| specific language governing permissions and limitations | ||||
| under the License. | ||||
| --> | ||||
| <project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" | ||||
| 	xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd"> | ||||
| 	<modelVersion>4.0.0</modelVersion> | ||||
|  | ||||
| 	<groupId>com.heibaiying</groupId> | ||||
| 	<artifactId>flink-kafka-integration</artifactId> | ||||
| 	<version>1.0</version> | ||||
| 	<packaging>jar</packaging> | ||||
|  | ||||
| 	<name>Flink Quickstart Job</name> | ||||
| 	<url>http://www.myorganization.org</url> | ||||
|  | ||||
| 	<properties> | ||||
| 		<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding> | ||||
| 		<flink.version>1.9.0</flink.version> | ||||
| 		<java.version>1.8</java.version> | ||||
| 		<scala.binary.version>2.11</scala.binary.version> | ||||
| 		<maven.compiler.source>${java.version}</maven.compiler.source> | ||||
| 		<maven.compiler.target>${java.version}</maven.compiler.target> | ||||
| 	</properties> | ||||
|  | ||||
| 	<repositories> | ||||
| 		<repository> | ||||
| 			<id>apache.snapshots</id> | ||||
| 			<name>Apache Development Snapshot Repository</name> | ||||
| 			<url>https://repository.apache.org/content/repositories/snapshots/</url> | ||||
| 			<releases> | ||||
| 				<enabled>false</enabled> | ||||
| 			</releases> | ||||
| 			<snapshots> | ||||
| 				<enabled>true</enabled> | ||||
| 			</snapshots> | ||||
| 		</repository> | ||||
| 	</repositories> | ||||
|  | ||||
| 	<dependencies> | ||||
| 		<!-- Apache Flink dependencies --> | ||||
| 		<!-- These dependencies are provided, because they should not be packaged into the JAR file. --> | ||||
| 		<dependency> | ||||
| 			<groupId>org.apache.flink</groupId> | ||||
| 			<artifactId>flink-java</artifactId> | ||||
| 			<version>${flink.version}</version> | ||||
| 			<scope>provided</scope> | ||||
| 		</dependency> | ||||
| 		<dependency> | ||||
| 			<groupId>org.apache.flink</groupId> | ||||
| 			<artifactId>flink-streaming-java_${scala.binary.version}</artifactId> | ||||
| 			<version>${flink.version}</version> | ||||
| 			<scope>provided</scope> | ||||
| 		</dependency> | ||||
|  | ||||
| 		<!-- Add connector dependencies here. They must be in the default scope (compile). --> | ||||
|  | ||||
| 		<!-- Example: | ||||
|  | ||||
| 		<dependency> | ||||
| 			<groupId>org.apache.flink</groupId> | ||||
| 			<artifactId>flink-connector-kafka-0.10_${scala.binary.version}</artifactId> | ||||
| 			<version>${flink.version}</version> | ||||
| 		</dependency> | ||||
| 		--> | ||||
|  | ||||
| 		<!-- Add logging framework, to produce console output when running in the IDE. --> | ||||
| 		<!-- These dependencies are excluded from the application JAR by default. --> | ||||
| 		<dependency> | ||||
| 			<groupId>org.slf4j</groupId> | ||||
| 			<artifactId>slf4j-log4j12</artifactId> | ||||
| 			<version>1.7.7</version> | ||||
| 			<scope>runtime</scope> | ||||
| 		</dependency> | ||||
| 		<dependency> | ||||
| 			<groupId>log4j</groupId> | ||||
| 			<artifactId>log4j</artifactId> | ||||
| 			<version>1.2.17</version> | ||||
| 			<scope>runtime</scope> | ||||
| 		</dependency> | ||||
| 		<dependency> | ||||
| 			<groupId>org.apache.flink</groupId> | ||||
| 			<artifactId>flink-connector-kafka_2.11</artifactId> | ||||
| 			<version>1.9.0</version> | ||||
| 		</dependency> | ||||
| 		<dependency> | ||||
| 			<groupId>mysql</groupId> | ||||
| 			<artifactId>mysql-connector-java</artifactId> | ||||
| 			<version>8.0.16</version> | ||||
| 		</dependency> | ||||
| 	</dependencies> | ||||
|  | ||||
| 	<build> | ||||
| 		<plugins> | ||||
|  | ||||
| 			<!-- Java Compiler --> | ||||
| 			<plugin> | ||||
| 				<groupId>org.apache.maven.plugins</groupId> | ||||
| 				<artifactId>maven-compiler-plugin</artifactId> | ||||
| 				<version>3.1</version> | ||||
| 				<configuration> | ||||
| 					<source>${java.version}</source> | ||||
| 					<target>${java.version}</target> | ||||
| 				</configuration> | ||||
| 			</plugin> | ||||
|  | ||||
| 			<!-- We use the maven-shade plugin to create a fat jar that contains all necessary dependencies. --> | ||||
| 			<!-- Change the value of <mainClass>...</mainClass> if your program entry point changes. --> | ||||
| 			<plugin> | ||||
| 				<groupId>org.apache.maven.plugins</groupId> | ||||
| 				<artifactId>maven-shade-plugin</artifactId> | ||||
| 				<version>3.0.0</version> | ||||
| 				<executions> | ||||
| 					<!-- Run shade goal on package phase --> | ||||
| 					<execution> | ||||
| 						<phase>package</phase> | ||||
| 						<goals> | ||||
| 							<goal>shade</goal> | ||||
| 						</goals> | ||||
| 						<configuration> | ||||
| 							<artifactSet> | ||||
| 								<excludes> | ||||
| 									<exclude>org.apache.flink:force-shading</exclude> | ||||
| 									<exclude>com.google.code.findbugs:jsr305</exclude> | ||||
| 									<exclude>org.slf4j:*</exclude> | ||||
| 									<exclude>log4j:*</exclude> | ||||
| 								</excludes> | ||||
| 							</artifactSet> | ||||
| 							<filters> | ||||
| 								<filter> | ||||
| 									<!-- Do not copy the signatures in the META-INF folder. | ||||
| 									Otherwise, this might cause SecurityExceptions when using the JAR. --> | ||||
| 									<artifact>*:*</artifact> | ||||
| 									<excludes> | ||||
| 										<exclude>META-INF/*.SF</exclude> | ||||
| 										<exclude>META-INF/*.DSA</exclude> | ||||
| 										<exclude>META-INF/*.RSA</exclude> | ||||
| 									</excludes> | ||||
| 								</filter> | ||||
| 							</filters> | ||||
| 							<transformers> | ||||
| 								<transformer implementation="org.apache.maven.plugins.shade.resource.ManifestResourceTransformer"> | ||||
| 									<mainClass>com.heibaiying.KafkaStreamingJob</mainClass> | ||||
| 								</transformer> | ||||
| 							</transformers> | ||||
| 						</configuration> | ||||
| 					</execution> | ||||
| 				</executions> | ||||
| 			</plugin> | ||||
| 		</plugins> | ||||
|  | ||||
| 		<pluginManagement> | ||||
| 			<plugins> | ||||
|  | ||||
| 				<!-- This improves the out-of-the-box experience in Eclipse by resolving some warnings. --> | ||||
| 				<plugin> | ||||
| 					<groupId>org.eclipse.m2e</groupId> | ||||
| 					<artifactId>lifecycle-mapping</artifactId> | ||||
| 					<version>1.0.0</version> | ||||
| 					<configuration> | ||||
| 						<lifecycleMappingMetadata> | ||||
| 							<pluginExecutions> | ||||
| 								<pluginExecution> | ||||
| 									<pluginExecutionFilter> | ||||
| 										<groupId>org.apache.maven.plugins</groupId> | ||||
| 										<artifactId>maven-shade-plugin</artifactId> | ||||
| 										<versionRange>[3.0.0,)</versionRange> | ||||
| 										<goals> | ||||
| 											<goal>shade</goal> | ||||
| 										</goals> | ||||
| 									</pluginExecutionFilter> | ||||
| 									<action> | ||||
| 										<ignore/> | ||||
| 									</action> | ||||
| 								</pluginExecution> | ||||
| 								<pluginExecution> | ||||
| 									<pluginExecutionFilter> | ||||
| 										<groupId>org.apache.maven.plugins</groupId> | ||||
| 										<artifactId>maven-compiler-plugin</artifactId> | ||||
| 										<versionRange>[3.1,)</versionRange> | ||||
| 										<goals> | ||||
| 											<goal>testCompile</goal> | ||||
| 											<goal>compile</goal> | ||||
| 										</goals> | ||||
| 									</pluginExecutionFilter> | ||||
| 									<action> | ||||
| 										<ignore/> | ||||
| 									</action> | ||||
| 								</pluginExecution> | ||||
| 							</pluginExecutions> | ||||
| 						</lifecycleMappingMetadata> | ||||
| 					</configuration> | ||||
| 				</plugin> | ||||
| 			</plugins> | ||||
| 		</pluginManagement> | ||||
| 	</build> | ||||
|  | ||||
| 	<!-- This profile helps to make things run out of the box in IntelliJ --> | ||||
| 	<!-- Its adds Flink's core classes to the runtime class path. --> | ||||
| 	<!-- Otherwise they are missing in IntelliJ, because the dependency is 'provided' --> | ||||
| 	<profiles> | ||||
| 		<profile> | ||||
| 			<id>add-dependencies-for-IDEA</id> | ||||
|  | ||||
| 			<activation> | ||||
| 				<property> | ||||
| 					<name>idea.version</name> | ||||
| 				</property> | ||||
| 			</activation> | ||||
|  | ||||
| 			<dependencies> | ||||
| 				<dependency> | ||||
| 					<groupId>org.apache.flink</groupId> | ||||
| 					<artifactId>flink-java</artifactId> | ||||
| 					<version>${flink.version}</version> | ||||
| 					<scope>compile</scope> | ||||
| 				</dependency> | ||||
| 				<dependency> | ||||
| 					<groupId>org.apache.flink</groupId> | ||||
| 					<artifactId>flink-streaming-java_${scala.binary.version}</artifactId> | ||||
| 					<version>${flink.version}</version> | ||||
| 					<scope>compile</scope> | ||||
| 				</dependency> | ||||
| 			</dependencies> | ||||
| 		</profile> | ||||
| 	</profiles> | ||||
|  | ||||
| </project> | ||||
| @@ -0,0 +1,23 @@ | ||||
| package com.heibaiying; | ||||
|  | ||||
| import com.heibaiying.bean.Employee; | ||||
| import com.heibaiying.sink.FlinkToMySQL; | ||||
| import org.apache.flink.streaming.api.datastream.DataStreamSource; | ||||
| import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment; | ||||
|  | ||||
| import java.sql.Date; | ||||
|  | ||||
| public class CustomSinkJob { | ||||
|  | ||||
|     public static void main(String[] args) throws Exception { | ||||
|  | ||||
|         final StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment(); | ||||
|         Date date = new Date(System.currentTimeMillis()); | ||||
|         DataStreamSource<Employee> streamSource = env.fromElements( | ||||
|                 new Employee("hei", 10, date), | ||||
|                 new Employee("bai", 20, date), | ||||
|                 new Employee("ying", 30, date)); | ||||
|         streamSource.addSink(new FlinkToMySQL()); | ||||
|         env.execute(); | ||||
|     } | ||||
| } | ||||
| @@ -0,0 +1,43 @@ | ||||
| package com.heibaiying; | ||||
|  | ||||
| import org.apache.flink.api.common.functions.MapFunction; | ||||
| import org.apache.flink.api.common.serialization.SimpleStringSchema; | ||||
| import org.apache.flink.streaming.api.datastream.DataStream; | ||||
| import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment; | ||||
| import org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer; | ||||
| import org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer; | ||||
| import org.apache.flink.streaming.connectors.kafka.KafkaSerializationSchema; | ||||
| import org.apache.kafka.clients.producer.ProducerRecord; | ||||
|  | ||||
| import javax.annotation.Nullable; | ||||
| import java.util.Properties; | ||||
|  | ||||
| public class KafkaStreamingJob { | ||||
|  | ||||
|     public static void main(String[] args) throws Exception { | ||||
|  | ||||
|         final StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment(); | ||||
|  | ||||
|         // 1.指定Kafka的相关配置属性 | ||||
|         Properties properties = new Properties(); | ||||
|         properties.setProperty("bootstrap.servers", "192.168.200.229:9092"); | ||||
|  | ||||
|         // 2.接收Kafka上的数据 | ||||
|         DataStream<String> stream = env | ||||
|                 .addSource(new FlinkKafkaConsumer<>("flink-stream-in-topic", new SimpleStringSchema(), properties)); | ||||
|  | ||||
|         // 3.定义计算结果到 Kafka ProducerRecord 的转换 | ||||
|         KafkaSerializationSchema<String> kafkaSerializationSchema = new KafkaSerializationSchema<String>() { | ||||
|             @Override | ||||
|             public ProducerRecord<byte[], byte[]> serialize(String element, @Nullable Long timestamp) { | ||||
|                 return new ProducerRecord<>("flink-stream-out-topic", element.getBytes()); | ||||
|             } | ||||
|         }; | ||||
|         // 4. 定义Flink Kafka生产者 | ||||
|         FlinkKafkaProducer<String> kafkaProducer = new FlinkKafkaProducer<>("flink-stream-out-topic", | ||||
|                 kafkaSerializationSchema, properties, FlinkKafkaProducer.Semantic.AT_LEAST_ONCE, 5); | ||||
|         // 5. 将接收到输入元素*2后写出到Kafka | ||||
|         stream.map((MapFunction<String, String>) value -> value + value).addSink(kafkaProducer); | ||||
|         env.execute("Flink Streaming"); | ||||
|     } | ||||
| } | ||||
| @@ -0,0 +1,42 @@ | ||||
| package com.heibaiying.bean; | ||||
|  | ||||
| import java.sql.Date; | ||||
|  | ||||
| public class Employee { | ||||
|  | ||||
|     private String name; | ||||
|     private int age; | ||||
|     private Date birthday; | ||||
|  | ||||
|     Employee(){} | ||||
|  | ||||
|     public Employee(String name, int age, Date birthday) { | ||||
|         this.name = name; | ||||
|         this.age = age; | ||||
|         this.birthday = birthday; | ||||
|     } | ||||
|  | ||||
|     public String getName() { | ||||
|         return name; | ||||
|     } | ||||
|  | ||||
|     public void setName(String name) { | ||||
|         this.name = name; | ||||
|     } | ||||
|  | ||||
|     public int getAge() { | ||||
|         return age; | ||||
|     } | ||||
|  | ||||
|     public void setAge(int age) { | ||||
|         this.age = age; | ||||
|     } | ||||
|  | ||||
|     public Date getBirthday() { | ||||
|         return birthday; | ||||
|     } | ||||
|  | ||||
|     public void setBirthday(Date birthday) { | ||||
|         this.birthday = birthday; | ||||
|     } | ||||
| } | ||||
| @@ -0,0 +1,43 @@ | ||||
| package com.heibaiying.sink; | ||||
|  | ||||
| import com.heibaiying.bean.Employee; | ||||
| import org.apache.flink.configuration.Configuration; | ||||
| import org.apache.flink.streaming.api.functions.sink.RichSinkFunction; | ||||
|  | ||||
| import java.sql.Connection; | ||||
| import java.sql.DriverManager; | ||||
| import java.sql.PreparedStatement; | ||||
|  | ||||
| public class FlinkToMySQL extends RichSinkFunction<Employee> { | ||||
|  | ||||
|     private PreparedStatement stmt; | ||||
|     private Connection conn; | ||||
|  | ||||
|     @Override | ||||
|     public void open(Configuration parameters) throws Exception { | ||||
|         Class.forName("com.mysql.cj.jdbc.Driver"); | ||||
|         conn = DriverManager.getConnection("jdbc:mysql://192.168.200.229:3306/employees?characterEncoding=UTF-8&serverTimezone=UTC&useSSL=false", "root", "123456"); | ||||
|         String sql = "insert into emp(name, age, birthday) values(?, ?, ?)"; | ||||
|         stmt = conn.prepareStatement(sql); | ||||
|     } | ||||
|  | ||||
|     @Override | ||||
|     public void invoke(Employee value, Context context) throws Exception { | ||||
|         stmt.setString(1, value.getName()); | ||||
|         stmt.setInt(2, value.getAge()); | ||||
|         stmt.setDate(3, value.getBirthday()); | ||||
|         stmt.executeUpdate(); | ||||
|     } | ||||
|  | ||||
|     @Override | ||||
|     public void close() throws Exception { | ||||
|         super.close(); | ||||
|         if (stmt != null) { | ||||
|             stmt.close(); | ||||
|         } | ||||
|         if (conn != null) { | ||||
|             conn.close(); | ||||
|         } | ||||
|     } | ||||
|  | ||||
| } | ||||
| @@ -0,0 +1,23 @@ | ||||
| ################################################################################ | ||||
| #  Licensed to the Apache Software Foundation (ASF) under one | ||||
| #  or more contributor license agreements.  See the NOTICE file | ||||
| #  distributed with this work for additional information | ||||
| #  regarding copyright ownership.  The ASF licenses this file | ||||
| #  to you under the Apache License, Version 2.0 (the | ||||
| #  "License"); you may not use this file except in compliance | ||||
| #  with the License.  You may obtain a copy of the License at | ||||
| # | ||||
| #      http://www.apache.org/licenses/LICENSE-2.0 | ||||
| # | ||||
| #  Unless required by applicable law or agreed to in writing, software | ||||
| #  distributed under the License is distributed on an "AS IS" BASIS, | ||||
| #  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | ||||
| #  See the License for the specific language governing permissions and | ||||
| # limitations under the License. | ||||
| ################################################################################ | ||||
|  | ||||
| log4j.rootLogger=INFO, console | ||||
|  | ||||
| log4j.appender.console=org.apache.log4j.ConsoleAppender | ||||
| log4j.appender.console.layout=org.apache.log4j.PatternLayout | ||||
| log4j.appender.console.layout.ConversionPattern=%d{HH:mm:ss,SSS} %-5p %-60c %x - %m%n | ||||
		Reference in New Issue
	
	Block a user