Group By, Rank and aggregate spark data frame using pyspark

前端 未结 2 1096

I have a dataframe that looks like:

A     B    C
---------------
A1    B1   0.8
A1    B2   0.55
A1    B3   0.43

A2    B1   0.7
A2    B2   0.5
A2    B3   0.5         


        
2条回答
  •  清酒与你
    2020-12-24 08:25

    windowSpec = Window.partitionBy("col1").orderBy("col2")
    
    ranked = demand.withColumn("col_rank", row_number().over(windowSpec))
    
    ranked.show(1000)
    

    [![example][1]][1]

提交回复
热议问题