[英]sql With partition use in spark sql dataframe query
以下是假設您的數據框名稱為 tempTrades 的問題代碼:
import pyspark.sql.functions as F
from pyspark.sql import Window
win_temp = Window.partitionBy(F.col("[date]")).orderBy(F.col("TradedVolumSum").desc())
tempTrades.withColumn(
"rn",
F.row_number().over(win_temp)
).filter(
F.col("rn") == 1
)
聲明:本站的技術帖子網頁,遵循CC BY-SA 4.0協議,如果您需要轉載,請注明本站網址或者原文地址。任何問題請咨詢:yoyou2525@163.com.