How to avoid reloading of same data into hdfs at the time of any failure in pyspark

后端 未结 0 559
时光取名叫无心
时光取名叫无心 2020-12-14 03:25

I have a pyspark program to perform business mapping and load the data into two hive external tables partitioned based on month end date.

Requested Scenario:<

相关标签:
回答
  • 消灭零回复
提交回复
热议问题