Junk Spark output file on S3 with dollar signs
问题 I have a simple spark job that reads a file from s3, takes five and writes back in s3. What I see is that there is always additional file in s3, next to my output "directory", which is called output_$folder$. What is it? How I can prevent spark from creating it? Here is some code to show what I am doing... x = spark.sparkContext.textFile("s3n://.../0000_part_00") five = x.take(5) five = spark.sparkContext.parallelize(five) five.repartition(1).saveAsTextFile("s3n://prod.casumo.stu/dimensions