原文链接
import org.apache.spark.{SparkConf, SparkContext}
import scala.collection.mutable
object SparkTest {
def main(args: Array[String]): Unit = {
val conf = new SparkConf().setAppName("SparkTest").setMaster("local[2]")
val sc = new SparkContext(conf)
val fileRdd = sc.parallelize(Array(("imsi1","2018-07-29 11:22:23","zd-A"),("imsi2","2018-07-29 11:22:24","zd-A"),("imsi3","2018-07-29 11:22:25","zd-A")))
val result = mutable.Map.empty[String,String]
val resultBroadCast = sc.broadcast(result)
fileRdd.foreach(input=>{
val str = (input._1+"\t"+input._2+"\t"+input._3).toString;
resultBroadCast.value += (input._1.toString -> str)
})
resultBroadCast.value.foreach(println(_))
sc.stop()
}
}
网友评论