【发布时间】:2016-10-03 09:59:40
【问题描述】:
对于一组数据帧
val df1 = sc.parallelize(1 to 4).map(i => (i,i*10)).toDF("id","x")
val df2 = sc.parallelize(1 to 4).map(i => (i,i*100)).toDF("id","y")
val df3 = sc.parallelize(1 to 4).map(i => (i,i*1000)).toDF("id","z")
把他们都联合起来
df1.unionAll(df2).unionAll(df3)
对于任意数量的数据帧是否有更优雅和可扩展的方式来执行此操作,例如来自
Seq(df1, df2, df3)
【问题讨论】:
标签: scala apache-spark apache-spark-sql