【发布时间】:2018-10-27 06:11:01
【问题描述】:
我有一个包含每天得分的数据框,我想计算每个用户的累积运行得分。我需要在一个新列上将前一天的累积分数与今天的分数相加,我尝试了 lag 函数进行计算,但由于某些原因它给出了错误。
这是我试过的代码:
import org.apache.spark.sql.expressions.Window
import org.apache.spark.sql.functions._
val genre = sc.parallelize(List(("Alice", "2016-05-01", "action",0),
("Alice", "2016-05-02", "0",1),
("Alice", "2016-05-03", "comedy",0),
("Alice", "2016-05-04", "action",1),
("Alice", "2016-05-05", "action",0),
("Alice", "2016-05-06", "horror",1),
("Bob", "2016-05-01", "art",0),
("Bob", "2016-05-02", "0",1),
("Bob", "2016-05-03", "0",0),
("Bob", "2016-05-04", "art",0),
("Bob", "2016-05-05", "comedy",1),
("Bob", "2016-05-06", "action",0))).
toDF("name", "date", "genre","score")
val wSpec2 = Window.partitionBy("name","genre").orderBy("date").rowsBetween(Long.MinValue, 0)
genre.withColumn( "CumScore",genre("score")*2+ lag(("CumScore"),1).over(wSpec2)*2 ).show()
数据框:
-----+----------+------+-----+
| name| date| genre|score|
+-----+----------+------+-----+
|Alice|2016-05-01|action| 0|
|Alice|2016-05-02| 0| 1|
|Alice|2016-05-03|comedy| 0|
|Alice|2016-05-04|action| 1|
|Alice|2016-05-05|action| 0|
|Alice|2016-05-06|horror| 1|
| Bob|2016-05-01| art| 0|
| Bob|2016-05-02| 0| 1|
| Bob|2016-05-03| 0| 0|
| Bob|2016-05-04| art| 0|
| Bob|2016-05-05|comedy| 1|
| Bob|2016-05-06|action| 0|
+-----+----------+------+-----+
我遇到的错误
org.apache.spark.sql.AnalysisException: Window Frame specifiedwindowframe(RowFrame, unboundedpreceding$(), currentrow$()) must match the required frame specifiedwindowframe(RowFrame, -1, -1);
at org.apa
【问题讨论】:
标签: scala apache-spark apache-spark-sql cumulative-sum