how can you calculate the size of an apache spark data frame using pyspark?

☆樱花仙子☆ 提交于 2019-12-22 04:33:08

问题


Is there a way to calculate the size in bytes of an Apache spark Data Frame using pyspark?


回答1:


why don't you just cache the df and then look in the spark UI under storage and convert the units to bytes

df.cache()


来源:https://stackoverflow.com/questions/38180140/how-can-you-calculate-the-size-of-an-apache-spark-data-frame-using-pyspark

易学教程内所有资源均来自网络或用户发布的内容,如有违反法律规定的内容欢迎反馈
该文章没有解决你所遇到的问题?点击提问,说说你的问题,让更多的人一起探讨吧!