How to avoid empty files while writing parquet files?

后端 未结 4 920
囚心锁ツ
囚心锁ツ 2021-01-16 07:15

I am reading from Kafka queue using Spark Structured Streaming. After reading from Kafka I am applying filter on the dataframe. I am saving this fi

4条回答
  •  醉话见心
    2021-01-16 07:46

    you can try with repartitionByRange(column)..

    I used this while writing dataframe to HDFS .. It solved my empty file creation issue.

提交回复
热议问题