strom整合其他框架
This commit is contained in:
parent
a339956c1a
commit
d4daa775cf
@ -14,7 +14,7 @@ import java.util.List;
|
|||||||
|
|
||||||
public class HBaseUtilsTest {
|
public class HBaseUtilsTest {
|
||||||
|
|
||||||
private static final String TABLE_NAME = "class";
|
private static final String TABLE_NAME = "WordCount";
|
||||||
private static final String TEACHER = "teacher";
|
private static final String TEACHER = "teacher";
|
||||||
private static final String STUDENT = "student";
|
private static final String STUDENT = "student";
|
||||||
|
|
||||||
@ -67,7 +67,7 @@ public class HBaseUtilsTest {
|
|||||||
ResultScanner scanner = HBaseUtils.getScanner(TABLE_NAME);
|
ResultScanner scanner = HBaseUtils.getScanner(TABLE_NAME);
|
||||||
if (scanner != null) {
|
if (scanner != null) {
|
||||||
scanner.forEach(result -> System.out.println(Bytes.toString(result.getRow()) + "->" + Bytes
|
scanner.forEach(result -> System.out.println(Bytes.toString(result.getRow()) + "->" + Bytes
|
||||||
.toString(result.getValue(Bytes.toBytes(STUDENT), Bytes.toBytes("name")))));
|
.toString(result.getValue(Bytes.toBytes("cf"), Bytes.toBytes("count")))));
|
||||||
scanner.close();
|
scanner.close();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
57
code/Storm/storm-hbase-integration/pom.xml
Normal file
57
code/Storm/storm-hbase-integration/pom.xml
Normal file
@ -0,0 +1,57 @@
|
|||||||
|
<?xml version="1.0" encoding="UTF-8"?>
|
||||||
|
<project xmlns="http://maven.apache.org/POM/4.0.0"
|
||||||
|
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
|
||||||
|
xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd">
|
||||||
|
<modelVersion>4.0.0</modelVersion>
|
||||||
|
|
||||||
|
<groupId>com.heibaiying</groupId>
|
||||||
|
<artifactId>storm-hbase-integration</artifactId>
|
||||||
|
<version>1.0</version>
|
||||||
|
|
||||||
|
<properties>
|
||||||
|
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
|
||||||
|
<storm.version>1.2.2</storm.version>
|
||||||
|
</properties>
|
||||||
|
|
||||||
|
<build>
|
||||||
|
<plugins>
|
||||||
|
<plugin>
|
||||||
|
<groupId>org.apache.maven.plugins</groupId>
|
||||||
|
<artifactId>maven-compiler-plugin</artifactId>
|
||||||
|
<configuration>
|
||||||
|
<source>8</source>
|
||||||
|
<target>8</target>
|
||||||
|
</configuration>
|
||||||
|
</plugin>
|
||||||
|
<plugin>
|
||||||
|
<artifactId>maven-assembly-plugin</artifactId>
|
||||||
|
<configuration>
|
||||||
|
<descriptors>
|
||||||
|
<descriptor>src/main/resources/assembly.xml</descriptor>
|
||||||
|
</descriptors>
|
||||||
|
<archive>
|
||||||
|
<manifest>
|
||||||
|
<mainClass>com.heibaiying.wordcount.ClusterWordCountApp</mainClass>
|
||||||
|
</manifest>
|
||||||
|
</archive>
|
||||||
|
</configuration>
|
||||||
|
</plugin>
|
||||||
|
</plugins>
|
||||||
|
</build>
|
||||||
|
|
||||||
|
|
||||||
|
<dependencies>
|
||||||
|
<dependency>
|
||||||
|
<groupId>org.apache.storm</groupId>
|
||||||
|
<artifactId>storm-core</artifactId>
|
||||||
|
<version>${storm.version}</version>
|
||||||
|
</dependency>
|
||||||
|
<!--Storm整合HBase依赖-->
|
||||||
|
<dependency>
|
||||||
|
<groupId>org.apache.storm</groupId>
|
||||||
|
<artifactId>storm-hbase</artifactId>
|
||||||
|
<version>${storm.version}</version>
|
||||||
|
</dependency>
|
||||||
|
</dependencies>
|
||||||
|
|
||||||
|
</project>
|
@ -0,0 +1,78 @@
|
|||||||
|
package com.heibaiying;
|
||||||
|
|
||||||
|
import com.heibaiying.component.DataSourceSpout;
|
||||||
|
import com.heibaiying.component.SplitBolt;
|
||||||
|
import org.apache.storm.Config;
|
||||||
|
import org.apache.storm.LocalCluster;
|
||||||
|
import org.apache.storm.StormSubmitter;
|
||||||
|
import org.apache.storm.generated.AlreadyAliveException;
|
||||||
|
import org.apache.storm.generated.AuthorizationException;
|
||||||
|
import org.apache.storm.generated.InvalidTopologyException;
|
||||||
|
import org.apache.storm.hbase.bolt.HBaseBolt;
|
||||||
|
import org.apache.storm.hbase.bolt.mapper.SimpleHBaseMapper;
|
||||||
|
import org.apache.storm.topology.TopologyBuilder;
|
||||||
|
import org.apache.storm.tuple.Fields;
|
||||||
|
|
||||||
|
import java.util.HashMap;
|
||||||
|
import java.util.Map;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* 进行词频统计 并将统计结果存储到HBase中
|
||||||
|
* <p>
|
||||||
|
* 编译打包: mvn clean assembly:assembly -Dmaven.test.skip=true
|
||||||
|
* hdfs://hadoop001:8020/hbase
|
||||||
|
*/
|
||||||
|
public class WordCountToHBaseApp {
|
||||||
|
|
||||||
|
private static final String DATA_SOURCE_SPOUT = "dataSourceSpout";
|
||||||
|
private static final String SPLIT_BOLT = "splitBolt";
|
||||||
|
private static final String COUNT_BOLT = "countBolt";
|
||||||
|
private static final String HBASE_BOLT = "hbaseBolt";
|
||||||
|
|
||||||
|
public static void main(String[] args) {
|
||||||
|
|
||||||
|
// storm的配置
|
||||||
|
Config config = new Config();
|
||||||
|
|
||||||
|
// HBase的配置
|
||||||
|
Map<String, Object> hbConf = new HashMap<>();
|
||||||
|
hbConf.put("hbase.rootdir", "hdfs://hadoop001:8020/hbase");
|
||||||
|
hbConf.put("hbase.zookeeper.quorum", "hadoop001:2181");
|
||||||
|
|
||||||
|
// 将HBase的配置传入Storm的配置中
|
||||||
|
config.put("hbase.conf", hbConf);
|
||||||
|
|
||||||
|
// 定义流数据与HBase中数据的映射
|
||||||
|
SimpleHBaseMapper mapper = new SimpleHBaseMapper()
|
||||||
|
.withRowKeyField("word")
|
||||||
|
.withColumnFields(new Fields("word"))
|
||||||
|
.withCounterFields(new Fields("count"))
|
||||||
|
.withColumnFamily("cf");
|
||||||
|
|
||||||
|
// 给HBaseBolt传入表名、数据映射关系、和HBase的配置信息
|
||||||
|
HBaseBolt hbase = new HBaseBolt("WordCount", mapper)
|
||||||
|
.withConfigKey("hbase.conf");
|
||||||
|
|
||||||
|
// 构建Topology
|
||||||
|
TopologyBuilder builder = new TopologyBuilder();
|
||||||
|
builder.setSpout(DATA_SOURCE_SPOUT, new DataSourceSpout(),1);
|
||||||
|
// split
|
||||||
|
builder.setBolt(SPLIT_BOLT, new SplitBolt(), 1).shuffleGrouping(DATA_SOURCE_SPOUT);
|
||||||
|
// save to HBase
|
||||||
|
builder.setBolt(HBASE_BOLT, hbase, 1).fieldsGrouping(SPLIT_BOLT, new Fields("word"));
|
||||||
|
|
||||||
|
|
||||||
|
// 如果外部传参cluster则代表线上环境启动,否则代表本地启动
|
||||||
|
if (args.length > 1 && args[1].equals("cluster")) {
|
||||||
|
try {
|
||||||
|
StormSubmitter.submitTopology("ClusterWordCountToRedisApp", config, builder.createTopology());
|
||||||
|
} catch (AlreadyAliveException | InvalidTopologyException | AuthorizationException e) {
|
||||||
|
e.printStackTrace();
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
LocalCluster cluster = new LocalCluster();
|
||||||
|
cluster.submitTopology("LocalWordCountToRedisApp",
|
||||||
|
config, builder.createTopology());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
@ -0,0 +1,53 @@
|
|||||||
|
package com.heibaiying.component;
|
||||||
|
|
||||||
|
import org.apache.storm.shade.org.apache.commons.lang.StringUtils;
|
||||||
|
import org.apache.storm.spout.SpoutOutputCollector;
|
||||||
|
import org.apache.storm.task.TopologyContext;
|
||||||
|
import org.apache.storm.topology.OutputFieldsDeclarer;
|
||||||
|
import org.apache.storm.topology.base.BaseRichSpout;
|
||||||
|
import org.apache.storm.tuple.Fields;
|
||||||
|
import org.apache.storm.tuple.Values;
|
||||||
|
import org.apache.storm.utils.Utils;
|
||||||
|
|
||||||
|
import java.util.*;
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* 产生词频样本的数据源
|
||||||
|
*/
|
||||||
|
public class DataSourceSpout extends BaseRichSpout {
|
||||||
|
|
||||||
|
private List<String> list = Arrays.asList("Spark", "Hadoop", "HBase", "Storm", "Flink", "Hive");
|
||||||
|
|
||||||
|
private SpoutOutputCollector spoutOutputCollector;
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void open(Map map, TopologyContext topologyContext, SpoutOutputCollector spoutOutputCollector) {
|
||||||
|
this.spoutOutputCollector = spoutOutputCollector;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void nextTuple() {
|
||||||
|
// 模拟产生数据
|
||||||
|
String lineData = productData();
|
||||||
|
spoutOutputCollector.emit(new Values(lineData));
|
||||||
|
Utils.sleep(1000);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void declareOutputFields(OutputFieldsDeclarer outputFieldsDeclarer) {
|
||||||
|
outputFieldsDeclarer.declare(new Fields("line"));
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* 模拟数据
|
||||||
|
*/
|
||||||
|
private String productData() {
|
||||||
|
Collections.shuffle(list);
|
||||||
|
Random random = new Random();
|
||||||
|
int endIndex = random.nextInt(list.size()) % (list.size()) + 1;
|
||||||
|
return StringUtils.join(list.toArray(), "\t", 0, endIndex);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
@ -0,0 +1,40 @@
|
|||||||
|
package com.heibaiying.component;
|
||||||
|
|
||||||
|
import org.apache.storm.task.OutputCollector;
|
||||||
|
import org.apache.storm.task.TopologyContext;
|
||||||
|
import org.apache.storm.topology.OutputFieldsDeclarer;
|
||||||
|
import org.apache.storm.topology.base.BaseRichBolt;
|
||||||
|
import org.apache.storm.tuple.Fields;
|
||||||
|
import org.apache.storm.tuple.Tuple;
|
||||||
|
import org.apache.storm.tuple.Values;
|
||||||
|
|
||||||
|
import java.util.Map;
|
||||||
|
|
||||||
|
import static org.apache.storm.utils.Utils.tuple;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* 将每行数据按照指定分隔符进行拆分
|
||||||
|
*/
|
||||||
|
public class SplitBolt extends BaseRichBolt {
|
||||||
|
|
||||||
|
private OutputCollector collector;
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void prepare(Map stormConf, TopologyContext context, OutputCollector collector) {
|
||||||
|
this.collector = collector;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void execute(Tuple input) {
|
||||||
|
String line = input.getStringByField("line");
|
||||||
|
String[] words = line.split("\t");
|
||||||
|
for (String word : words) {
|
||||||
|
collector.emit(tuple(word, 1));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void declareOutputFields(OutputFieldsDeclarer declarer) {
|
||||||
|
declarer.declare(new Fields("word", "count"));
|
||||||
|
}
|
||||||
|
}
|
@ -0,0 +1,25 @@
|
|||||||
|
<assembly xmlns="http://maven.apache.org/ASSEMBLY/2.0.0"
|
||||||
|
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
|
||||||
|
xsi:schemaLocation="http://maven.apache.org/ASSEMBLY/2.0.0 http://maven.apache.org/xsd/assembly-2.0.0.xsd">
|
||||||
|
|
||||||
|
<id>with-dependencies</id>
|
||||||
|
|
||||||
|
<!--指明打包方式-->
|
||||||
|
<formats>
|
||||||
|
<format>jar</format>
|
||||||
|
</formats>
|
||||||
|
|
||||||
|
<includeBaseDirectory>false</includeBaseDirectory>
|
||||||
|
<dependencySets>
|
||||||
|
<dependencySet>
|
||||||
|
<outputDirectory>/</outputDirectory>
|
||||||
|
<useProjectArtifact>true</useProjectArtifact>
|
||||||
|
<unpack>true</unpack>
|
||||||
|
<scope>runtime</scope>
|
||||||
|
<!--排除storm环境中已经提供的storm-core-->
|
||||||
|
<excludes>
|
||||||
|
<exclude>org.apache.storm:storm-core</exclude>
|
||||||
|
</excludes>
|
||||||
|
</dependencySet>
|
||||||
|
</dependencySets>
|
||||||
|
</assembly>
|
110
code/Storm/storm-hdfs-integration/pom.xml
Normal file
110
code/Storm/storm-hdfs-integration/pom.xml
Normal file
@ -0,0 +1,110 @@
|
|||||||
|
<?xml version="1.0" encoding="UTF-8"?>
|
||||||
|
<project xmlns="http://maven.apache.org/POM/4.0.0"
|
||||||
|
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
|
||||||
|
xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd">
|
||||||
|
<modelVersion>4.0.0</modelVersion>
|
||||||
|
|
||||||
|
<groupId>com.heibaiying</groupId>
|
||||||
|
<artifactId>storm-hdfs-integration</artifactId>
|
||||||
|
<version>1.0</version>
|
||||||
|
|
||||||
|
<properties>
|
||||||
|
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
|
||||||
|
<storm.version>1.2.2</storm.version>
|
||||||
|
</properties>
|
||||||
|
|
||||||
|
<repositories>
|
||||||
|
<repository>
|
||||||
|
<id>cloudera</id>
|
||||||
|
<url>https://repository.cloudera.com/artifactory/cloudera-repos/</url>
|
||||||
|
</repository>
|
||||||
|
</repositories>
|
||||||
|
|
||||||
|
|
||||||
|
<build>
|
||||||
|
<plugins>
|
||||||
|
<!--使用java8编译-->
|
||||||
|
<plugin>
|
||||||
|
<groupId>org.apache.maven.plugins</groupId>
|
||||||
|
<artifactId>maven-compiler-plugin</artifactId>
|
||||||
|
<configuration>
|
||||||
|
<source>8</source>
|
||||||
|
<target>8</target>
|
||||||
|
</configuration>
|
||||||
|
</plugin>
|
||||||
|
|
||||||
|
<!--使用shade进行打包-->
|
||||||
|
<plugin>
|
||||||
|
<groupId>org.apache.maven.plugins</groupId>
|
||||||
|
<artifactId>maven-shade-plugin</artifactId>
|
||||||
|
<configuration>
|
||||||
|
<createDependencyReducedPom>true</createDependencyReducedPom>
|
||||||
|
<filters>
|
||||||
|
<filter>
|
||||||
|
<artifact>*:*</artifact>
|
||||||
|
<excludes>
|
||||||
|
<exclude>org.apache.storm:storm-core</exclude>
|
||||||
|
</excludes>
|
||||||
|
</filter>
|
||||||
|
</filters>
|
||||||
|
</configuration>
|
||||||
|
<executions>
|
||||||
|
<execution>
|
||||||
|
<phase>package</phase>
|
||||||
|
<goals>
|
||||||
|
<goal>shade</goal>
|
||||||
|
</goals>
|
||||||
|
<configuration>
|
||||||
|
<transformers>
|
||||||
|
<transformer
|
||||||
|
implementation="org.apache.maven.plugins.shade.resource.ServicesResourceTransformer"/>
|
||||||
|
<transformer
|
||||||
|
implementation="org.apache.maven.plugins.shade.resource.ManifestResourceTransformer">
|
||||||
|
</transformer>
|
||||||
|
</transformers>
|
||||||
|
</configuration>
|
||||||
|
</execution>
|
||||||
|
</executions>
|
||||||
|
</plugin>
|
||||||
|
</plugins>
|
||||||
|
</build>
|
||||||
|
|
||||||
|
|
||||||
|
<dependencies>
|
||||||
|
<dependency>
|
||||||
|
<groupId>org.apache.storm</groupId>
|
||||||
|
<artifactId>storm-core</artifactId>
|
||||||
|
<version>${storm.version}</version>
|
||||||
|
</dependency>
|
||||||
|
<!--Storm整合HDFS依赖-->
|
||||||
|
<dependency>
|
||||||
|
<groupId>org.apache.storm</groupId>
|
||||||
|
<artifactId>storm-hdfs</artifactId>
|
||||||
|
<version>${storm.version}</version>
|
||||||
|
</dependency>
|
||||||
|
<dependency>
|
||||||
|
<groupId>org.apache.hadoop</groupId>
|
||||||
|
<artifactId>hadoop-client</artifactId>
|
||||||
|
<version>2.6.0-cdh5.15.2</version>
|
||||||
|
<exclusions>
|
||||||
|
<exclusion>
|
||||||
|
<groupId>org.slf4j</groupId>
|
||||||
|
<artifactId>slf4j-log4j12</artifactId>
|
||||||
|
</exclusion>
|
||||||
|
</exclusions>
|
||||||
|
</dependency>
|
||||||
|
<dependency>
|
||||||
|
<groupId>org.apache.hadoop</groupId>
|
||||||
|
<artifactId>hadoop-hdfs</artifactId>
|
||||||
|
<version>2.6.0-cdh5.15.2</version>
|
||||||
|
<exclusions>
|
||||||
|
<exclusion>
|
||||||
|
<groupId>org.slf4j</groupId>
|
||||||
|
<artifactId>slf4j-log4j12</artifactId>
|
||||||
|
</exclusion>
|
||||||
|
</exclusions>
|
||||||
|
</dependency>
|
||||||
|
</dependencies>
|
||||||
|
|
||||||
|
|
||||||
|
</project>
|
@ -0,0 +1,77 @@
|
|||||||
|
package com.heibaiying;
|
||||||
|
|
||||||
|
import com.heibaiying.component.DataSourceSpout;
|
||||||
|
import org.apache.storm.Config;
|
||||||
|
import org.apache.storm.LocalCluster;
|
||||||
|
import org.apache.storm.StormSubmitter;
|
||||||
|
import org.apache.storm.generated.AlreadyAliveException;
|
||||||
|
import org.apache.storm.generated.AuthorizationException;
|
||||||
|
import org.apache.storm.generated.InvalidTopologyException;
|
||||||
|
import org.apache.storm.hdfs.bolt.HdfsBolt;
|
||||||
|
import org.apache.storm.hdfs.bolt.format.DefaultFileNameFormat;
|
||||||
|
import org.apache.storm.hdfs.bolt.format.DelimitedRecordFormat;
|
||||||
|
import org.apache.storm.hdfs.bolt.format.FileNameFormat;
|
||||||
|
import org.apache.storm.hdfs.bolt.format.RecordFormat;
|
||||||
|
import org.apache.storm.hdfs.bolt.rotation.FileRotationPolicy;
|
||||||
|
import org.apache.storm.hdfs.bolt.rotation.FileSizeRotationPolicy;
|
||||||
|
import org.apache.storm.hdfs.bolt.rotation.FileSizeRotationPolicy.Units;
|
||||||
|
import org.apache.storm.hdfs.bolt.sync.CountSyncPolicy;
|
||||||
|
import org.apache.storm.hdfs.bolt.sync.SyncPolicy;
|
||||||
|
import org.apache.storm.topology.TopologyBuilder;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* 进行词频统计 并将统计结果存储到HDFS中
|
||||||
|
* <p>
|
||||||
|
* hdfs://hadoopp001:8020 path
|
||||||
|
*/
|
||||||
|
public class WordCountToHdfsApp {
|
||||||
|
|
||||||
|
private static final String DATA_SOURCE_SPOUT = "dataSourceSpout";
|
||||||
|
private static final String HDFS_BOLT = "hdfsBolt";
|
||||||
|
|
||||||
|
public static void main(String[] args) {
|
||||||
|
|
||||||
|
// 定义存储文本的分隔符
|
||||||
|
RecordFormat format = new DelimitedRecordFormat()
|
||||||
|
.withFieldDelimiter("|");
|
||||||
|
|
||||||
|
// 同步策略: 每100个tuples之后就会把数据从缓存刷新到HDFS中
|
||||||
|
SyncPolicy syncPolicy = new CountSyncPolicy(100);
|
||||||
|
|
||||||
|
// 文件策略: 每个文件大小上限1M,超过限定时,创建新文件并继续写入
|
||||||
|
FileRotationPolicy rotationPolicy = new FileSizeRotationPolicy(1.0f, Units.MB);
|
||||||
|
|
||||||
|
// 定义完整路径
|
||||||
|
FileNameFormat fileNameFormat = new DefaultFileNameFormat()
|
||||||
|
.withPath("/storm-hdfs/");
|
||||||
|
|
||||||
|
// 定义HdfsBolt
|
||||||
|
HdfsBolt hdfsBolt = new HdfsBolt()
|
||||||
|
.withFsUrl("hdfs://hadoop001:8020")
|
||||||
|
.withFileNameFormat(fileNameFormat)
|
||||||
|
.withRecordFormat(format)
|
||||||
|
.withRotationPolicy(rotationPolicy)
|
||||||
|
.withSyncPolicy(syncPolicy);
|
||||||
|
|
||||||
|
|
||||||
|
// 构建Topology
|
||||||
|
TopologyBuilder builder = new TopologyBuilder();
|
||||||
|
builder.setSpout(DATA_SOURCE_SPOUT, new DataSourceSpout());
|
||||||
|
// save to HBase
|
||||||
|
builder.setBolt(HDFS_BOLT, hdfsBolt, 1).shuffleGrouping(DATA_SOURCE_SPOUT);
|
||||||
|
|
||||||
|
|
||||||
|
// 如果外部传参cluster则代表线上环境启动,否则代表本地启动
|
||||||
|
if (args.length > 0 && args[0].equals("cluster")) {
|
||||||
|
try {
|
||||||
|
StormSubmitter.submitTopology("ClusterWordCountToHdfsApp", new Config(), builder.createTopology());
|
||||||
|
} catch (AlreadyAliveException | InvalidTopologyException | AuthorizationException e) {
|
||||||
|
e.printStackTrace();
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
LocalCluster cluster = new LocalCluster();
|
||||||
|
cluster.submitTopology("LocalWordCountToHdfsApp",
|
||||||
|
new Config(), builder.createTopology());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
@ -0,0 +1,53 @@
|
|||||||
|
package com.heibaiying.component;
|
||||||
|
|
||||||
|
import org.apache.storm.shade.org.apache.commons.lang.StringUtils;
|
||||||
|
import org.apache.storm.spout.SpoutOutputCollector;
|
||||||
|
import org.apache.storm.task.TopologyContext;
|
||||||
|
import org.apache.storm.topology.OutputFieldsDeclarer;
|
||||||
|
import org.apache.storm.topology.base.BaseRichSpout;
|
||||||
|
import org.apache.storm.tuple.Fields;
|
||||||
|
import org.apache.storm.tuple.Values;
|
||||||
|
import org.apache.storm.utils.Utils;
|
||||||
|
|
||||||
|
import java.util.*;
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* 产生词频样本的数据源
|
||||||
|
*/
|
||||||
|
public class DataSourceSpout extends BaseRichSpout {
|
||||||
|
|
||||||
|
private List<String> list = Arrays.asList("Spark", "Hadoop", "HBase", "Storm", "Flink", "Hive");
|
||||||
|
|
||||||
|
private SpoutOutputCollector spoutOutputCollector;
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void open(Map map, TopologyContext topologyContext, SpoutOutputCollector spoutOutputCollector) {
|
||||||
|
this.spoutOutputCollector = spoutOutputCollector;
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void nextTuple() {
|
||||||
|
// 模拟产生数据
|
||||||
|
String lineData = productData();
|
||||||
|
spoutOutputCollector.emit(new Values(lineData));
|
||||||
|
Utils.sleep(1000);
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public void declareOutputFields(OutputFieldsDeclarer outputFieldsDeclarer) {
|
||||||
|
outputFieldsDeclarer.declare(new Fields("line"));
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* 模拟数据
|
||||||
|
*/
|
||||||
|
private String productData() {
|
||||||
|
Collections.shuffle(list);
|
||||||
|
Random random = new Random();
|
||||||
|
int endIndex = random.nextInt(list.size()) % (list.size()) + 1;
|
||||||
|
return StringUtils.join(list.toArray(), "\t", 0, endIndex);
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
@ -10,7 +10,7 @@
|
|||||||
|
|
||||||
<properties>
|
<properties>
|
||||||
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
|
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
|
||||||
<storm.version>1.2.0</storm.version>
|
<storm.version>1.2.2</storm.version>
|
||||||
</properties>
|
</properties>
|
||||||
|
|
||||||
<build>
|
<build>
|
||||||
|
@ -18,22 +18,26 @@ import org.apache.storm.topology.TopologyBuilder;
|
|||||||
*/
|
*/
|
||||||
public class CustomRedisCountApp {
|
public class CustomRedisCountApp {
|
||||||
|
|
||||||
|
private static final String DATA_SOURCE_SPOUT = "dataSourceSpout";
|
||||||
|
private static final String SPLIT_BOLT = "splitBolt";
|
||||||
|
private static final String STORE_BOLT = "storeBolt";
|
||||||
|
|
||||||
private static final String REDIS_HOST = "192.168.200.226";
|
private static final String REDIS_HOST = "192.168.200.226";
|
||||||
private static final int REDIS_PORT = 6379;
|
private static final int REDIS_PORT = 6379;
|
||||||
|
|
||||||
public static void main(String[] args) {
|
public static void main(String[] args) {
|
||||||
TopologyBuilder builder = new TopologyBuilder();
|
TopologyBuilder builder = new TopologyBuilder();
|
||||||
builder.setSpout("dataSourceSpout", new DataSourceSpout());
|
builder.setSpout(DATA_SOURCE_SPOUT, new DataSourceSpout());
|
||||||
// split
|
// split
|
||||||
builder.setBolt("splitBolt", new SplitBolt()).shuffleGrouping("dataSourceSpout");
|
builder.setBolt(SPLIT_BOLT, new SplitBolt()).shuffleGrouping(DATA_SOURCE_SPOUT);
|
||||||
// save to redis and count
|
// save to redis and count
|
||||||
JedisPoolConfig poolConfig = new JedisPoolConfig.Builder()
|
JedisPoolConfig poolConfig = new JedisPoolConfig.Builder()
|
||||||
.setHost(REDIS_HOST).setPort(REDIS_PORT).build();
|
.setHost(REDIS_HOST).setPort(REDIS_PORT).build();
|
||||||
RedisStoreMapper storeMapper = new WordCountStoreMapper();
|
RedisStoreMapper storeMapper = new WordCountStoreMapper();
|
||||||
RedisCountStoreBolt countStoreBolt = new RedisCountStoreBolt(poolConfig, storeMapper);
|
RedisCountStoreBolt countStoreBolt = new RedisCountStoreBolt(poolConfig, storeMapper);
|
||||||
builder.setBolt("storeBolt", countStoreBolt).shuffleGrouping("splitBolt");
|
builder.setBolt(STORE_BOLT, countStoreBolt).shuffleGrouping(SPLIT_BOLT);
|
||||||
|
|
||||||
// 如果外部传参cluster则代表线上环境启动否则代表本地启动
|
// 如果外部传参cluster则代表线上环境启动,否则代表本地启动
|
||||||
if (args.length > 0 && args[0].equals("cluster")) {
|
if (args.length > 0 && args[0].equals("cluster")) {
|
||||||
try {
|
try {
|
||||||
StormSubmitter.submitTopology("ClusterCustomRedisCountApp", new Config(), builder.createTopology());
|
StormSubmitter.submitTopology("ClusterCustomRedisCountApp", new Config(), builder.createTopology());
|
||||||
|
@ -17,29 +17,35 @@ import org.apache.storm.topology.TopologyBuilder;
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* 进行词频统计 并将统计结果存储到Redis中
|
* 进行词频统计 并将统计结果存储到Redis中
|
||||||
*
|
* <p>
|
||||||
* 编译打包: mvn clean assembly:assembly -Dmaven.test.skip=true
|
* 编译打包: mvn clean assembly:assembly -Dmaven.test.skip=true
|
||||||
* 提交Topology到集群: storm jar /usr/appjar/storm-redis-integration-1.0-jar-with-dependencies.jar com.heibaiying.WordCountToRedisApp cluster
|
* 提交Topology到集群: storm jar /usr/appjar/storm-redis-integration-1.0-with-dependencies.jar com.heibaiying.WordCountToRedisApp cluster
|
||||||
* 停止Topology: storm kill ClusterWordCountApp -w 3
|
* 停止Topology: storm kill ClusterWordCountApp -w 3
|
||||||
*/
|
*/
|
||||||
public class WordCountToRedisApp {
|
public class WordCountToRedisApp {
|
||||||
|
|
||||||
|
private static final String DATA_SOURCE_SPOUT = "dataSourceSpout";
|
||||||
|
private static final String SPLIT_BOLT = "splitBolt";
|
||||||
|
private static final String COUNT_BOLT = "countBolt";
|
||||||
|
private static final String STORE_BOLT = "storeBolt";
|
||||||
|
|
||||||
|
//在实际开发中这些参数可以将通过外部传入 使得程序更加灵活
|
||||||
private static final String REDIS_HOST = "192.168.200.226";
|
private static final String REDIS_HOST = "192.168.200.226";
|
||||||
private static final int REDIS_PORT = 6379;
|
private static final int REDIS_PORT = 6379;
|
||||||
|
|
||||||
public static void main(String[] args) {
|
public static void main(String[] args) {
|
||||||
TopologyBuilder builder = new TopologyBuilder();
|
TopologyBuilder builder = new TopologyBuilder();
|
||||||
builder.setSpout("dataSourceSpout", new DataSourceSpout());
|
builder.setSpout(DATA_SOURCE_SPOUT, new DataSourceSpout());
|
||||||
// split
|
// split
|
||||||
builder.setBolt("splitBolt", new SplitBolt()).shuffleGrouping("dataSourceSpout");
|
builder.setBolt(SPLIT_BOLT, new SplitBolt()).shuffleGrouping(DATA_SOURCE_SPOUT);
|
||||||
// count
|
// count
|
||||||
builder.setBolt("countBolt", new CountBolt()).shuffleGrouping("splitBolt");
|
builder.setBolt(COUNT_BOLT, new CountBolt()).shuffleGrouping(SPLIT_BOLT);
|
||||||
// save to redis
|
// save to redis
|
||||||
JedisPoolConfig poolConfig = new JedisPoolConfig.Builder()
|
JedisPoolConfig poolConfig = new JedisPoolConfig.Builder()
|
||||||
.setHost(REDIS_HOST).setPort(REDIS_PORT).build();
|
.setHost(REDIS_HOST).setPort(REDIS_PORT).build();
|
||||||
RedisStoreMapper storeMapper = new WordCountStoreMapper();
|
RedisStoreMapper storeMapper = new WordCountStoreMapper();
|
||||||
RedisStoreBolt storeBolt = new RedisStoreBolt(poolConfig, storeMapper);
|
RedisStoreBolt storeBolt = new RedisStoreBolt(poolConfig, storeMapper);
|
||||||
builder.setBolt("storeBolt", storeBolt).shuffleGrouping("countBolt");
|
builder.setBolt(STORE_BOLT, storeBolt).shuffleGrouping(COUNT_BOLT);
|
||||||
|
|
||||||
// 如果外部传参cluster则代表线上环境启动否则代表本地启动
|
// 如果外部传参cluster则代表线上环境启动否则代表本地启动
|
||||||
if (args.length > 0 && args[0].equals("cluster")) {
|
if (args.length > 0 && args[0].equals("cluster")) {
|
||||||
|
@ -2,7 +2,7 @@
|
|||||||
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
|
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
|
||||||
xsi:schemaLocation="http://maven.apache.org/ASSEMBLY/2.0.0 http://maven.apache.org/xsd/assembly-2.0.0.xsd">
|
xsi:schemaLocation="http://maven.apache.org/ASSEMBLY/2.0.0 http://maven.apache.org/xsd/assembly-2.0.0.xsd">
|
||||||
|
|
||||||
<id>jar-with-dependencies</id>
|
<id>with-dependencies</id>
|
||||||
|
|
||||||
<!--指明打包方式-->
|
<!--指明打包方式-->
|
||||||
<formats>
|
<formats>
|
||||||
|
Loading…
x
Reference in New Issue
Block a user