1. package tcode.day05
    2. import org.apache.spark.{SparkConf, SparkContext}
    3. import org.junit.Test
    4. class $07_ReadWrite {
    5. val sc = new SparkContext(new SparkConf().setMaster("local[4]").setAppName("test"))
    6. @Test
    7. def read(): Unit ={
    8. //读取文本数据
    9. sc.textFile("datas/wc.txt")
    10. //读取序列化文件
    11. val rdd = sc.sequenceFile[Int,Int]("output/seq")
    12. //println(rdd.collect().toList)
    13. //读取对象文件
    14. println(sc.objectFile[Int]("output/obj").collect().toList)
    15. }
    16. @Test
    17. def write(): Unit ={
    18. val rdd = sc.parallelize(List(1,2,3,5,5,6,7))
    19. //保存为文本
    20. //rdd.saveAsTextFile("output/text")
    21. //保存为对象文件
    22. //rdd.saveAsObjectFile("output/obj")
    23. val rdd2 = rdd.map(x=>(x,x))
    24. rdd2.saveAsSequenceFile("output/seq")
    25. //rdd2.saveAsNewAPIHadoopFile()
    26. //rdd2.saveAsHadoopFile()
    27. }
    28. }