Spark: Sort records in groups?

后端 未结 4 1176
忘掉有多难
忘掉有多难 2020-12-31 11:34

I have a set of records which I need to:

1) Group by \'date\', \'city\' and \'kind\'

2) Sort every group by \'prize

In my code:

impor         


        
4条回答
  •  温柔的废话
    2020-12-31 12:14

    Replace map with flatMap

    val x = rsGrp.map{r => 
      val lst = r.toList
      lst.map{e => (e.prize, e)}
      }
    

    this will give you a

    org.apache.spark.rdd.RDD[(Int, Record)] = FlatMappedRDD[10]
    

    and then you can call sortBy(_._1) on the RDD above.

提交回复
热议问题