This commit is contained in:
罗祥 2019-05-15 08:52:35 +08:00
parent beb52978b7
commit a94cac32e6
9 changed files with 28 additions and 13 deletions

View File

@ -70,10 +70,28 @@
## 三、Spark
**Spark Core :**
1. Spark简介
2. [Spark单机版本环境搭建](https://github.com/heibaiying/BigData-Notes/blob/master/notes/installation/Spark%E5%8D%95%E6%9C%BA%E7%89%88%E6%9C%AC%E7%8E%AF%E5%A2%83%E6%90%AD%E5%BB%BA.md)
3. RDD详解
4. Spark Transformation 和 Action
3. 弹性式数据集RDD
4. Spark命令行的基本使用
5. RDD常用算子之——Transformation Action
6. RDD常用算子之——Action
7. Spark广播变量与累加器
**Spark SQL :**
1. DataFrame和DateSet
2. Spark SQL之常用SQL语句
3. External Data Source
**Spark Streaming **
1. Spark Streaming简介
2. DStream常用函数
3. Spark Streaming 整合 flume
4. Spark Streaming 整合 kafka
## 四、Flink

View File

@ -1 +0,0 @@
(mapreduce,1)

View File

@ -1,2 +0,0 @@
(spark,2)
(hadoop,2)

View File

@ -3,14 +3,12 @@ package com.heibaiying.spark.rdd
import org.apache.spark.{SparkConf, SparkContext}
object WordCount {
object WordCount extends App {
def main(args: Array[String]): Unit = {
val conf = new SparkConf().setAppName("sparkBase").setMaster("local[2]")
val sc = new SparkContext(conf)
val rdd = sc.textFile("input/wc.txt").flatMap(_.split(",")).map((_, 1)).reduceByKey(_ + _)
rdd.foreach(println)
rdd.saveAsTextFile("output/")
}
val conf = new SparkConf().setAppName("sparkBase").setMaster("local[2]")
val sc = new SparkContext(conf)
val rdd = sc.textFile("input/wc.txt").flatMap(_.split(",")).map((_, 1)).reduceByKey(_ + _)
rdd.foreach(println)
rdd.saveAsTextFile("output/")
}

2
notes/Spark-RDD.md Normal file
View File

@ -0,0 +1,2 @@
弹性式数据集RDD