Spark - How to write a single csv file WITHOUT folder?

前端 未结 9 1200
北恋
北恋 2020-12-28 13:44

Suppose that df is a dataframe in Spark. The way to write df into a single CSV file is

df.coalesce(1).write.option(\"header\", \"tru

9条回答
  •  轻奢々
    轻奢々 (楼主)
    2020-12-28 14:17

    A more databricks'y' solution is here:

    TEMPORARY_TARGET="dbfs:/my_folder/filename"
    DESIRED_TARGET="dbfs:/my_folder/filename.csv"
    
    spark_df.coalesce(1).write.option("header", "true").csv(TEMPORARY_TARGET)
    
    temporary_csv = os.path.join(TEMPORARY_TARGET, dbutils.fs.ls(TEMPORARY_TARGET)[3][1])
    
    dbutils.fs.cp(temporary_csv, DESIRED_TARGET)
    

    Note if you are working from Koalas data frame you can replace spark_df with koalas_df.to_spark()

提交回复
热议问题