Viewing internal Spark Dataframe contents

孤街醉人 提交于 2019-12-10 15:52:32

问题


When debugging a spark program, I can pause the stack and look at the frame to see all the meta data of a DataFrame. Partition metadata like input split, logical plan metadata, underlying RDD metadata, etc. But I cannot see the contents of the DataFrame. The DataFrame is another JVM somewhere on another node, or even on the same node (on a local training cluster). So my question, does anyone use a way for troubleshooting, where they are looking at the contents of the DataFrame partitions the way the driver program can be debugged?

来源:https://stackoverflow.com/questions/45243388/viewing-internal-spark-dataframe-contents

易学教程内所有资源均来自网络或用户发布的内容,如有违反法律规定的内容欢迎反馈
该文章没有解决你所遇到的问题?点击提问,说说你的问题,让更多的人一起探讨吧!