Pandas, large data, HDF tables and memory usage when calling a function
问题 Short question When Pandas work on a HDFStore (eg: .mean() or .apply() ), does it load the full data in memory as a DataFrame, or does it process record-by-record as a Serie? Long description I have to work on large data files, and I can specify the output format of the data file. I intend to use Pandas to process the data, and I would like to setup the best format so that it maximizes the performances. I have seen that panda.read_table() has gone a long way, but it still at least takes at