美文网首页
利用文件模拟数据表

利用文件模拟数据表

作者: Jorvi | 来源:发表于2021-07-10 11:21 被阅读0次
import org.apache.spark.SparkConf
import org.apache.spark.sql.SparkSession

object SparkApp {

  def main(args: Array[String]): Unit = {
    val conf = new SparkConf().setAppName("Spark Sql Test").setMaster("local[2]")
    val session = SparkSession.builder().config(conf).getOrCreate()

    import session.implicits._

    val sqlContext = session.sqlContext

    val people = sqlContext.read.textFile("D:\\resources\\test.txt")

    val peopleDs = people
      .map(_.split(","))
      .map(p => Person(p(0), p(1).trim.toInt, p(2)))

    peopleDs.createOrReplaceTempView("peopleTempView")

    val count = session.sql("SELECT count(1) FROM peopleTempView").rdd.first().getAs[Long](0)

    println(count)
  }

}


case class Person(name:String, age:Int, addr:String)

aa,25,chengdu
bb,23,beijing
cc,16,hangzhou
dd,19,zhejiang

相关文章

网友评论

      本文标题:利用文件模拟数据表

      本文链接:https://www.haomeiwen.com/subject/nlrcpltx.html