Pickling pandas dataframe multiplies by 5 the file size
问题 I am reading a 800 Mb CSV file with pandas.read_csv , and then use the original Python pickle.dump(datfarame) to save it. The result is a 4 Gb pkl file, so the CSV size is multiplied by 5. I expected pickle to compress data rather than extend it. Also because I can do a gzip on the CSV file which compress it to 200 Mb, dividing it by 4. I am willing to accelerate the loading time of my program, and thought that pickling would help, but considering disk access is the main bottleneck I am