Comparing columns in Pyspark

前端 未结 5 858
Happy的楠姐
Happy的楠姐 2020-12-01 18:17

I am working on a PySpark DataFrame with n columns. I have a set of m columns (m < n) and my task is choose the column with max values in it.

For example:

5条回答
  •  慢半拍i
    慢半拍i (楼主)
    2020-12-01 19:07

    You can also use the pyspark built-in least:

    from pyspark.sql.functions import least, col
    df = df.withColumn('min', least(col('c1'), col('c2'), col('c3')))
    

提交回复
热议问题