I have a large (about 85 GB compressed) gzipped file from s3 that I am trying to process with Spark on AWS EMR (right now with an m4.xlarge master instance and two m4.10xlar
Spark can parallelize reading a single gzip file.
The best you can do split it in chunks that are gzipped.
However, Spark is really slow at reading gzip files. You can do this to speed it up:
file_names_rdd = sc.parallelize(list_of_files, 100)
lines_rdd = file_names_rdd.flatMap(lambda _: gzip.open(_).readlines())
Going through Python is twice has fast as reading the native Spark gzip reader.