Pyspark socket timeout exception after application running for a while

后端 未结 3 1805
孤独总比滥情好
孤独总比滥情好 2021-01-05 08:32

I am using pyspark to estimate parameters for a logistic regression model. I use spark to calculate the likelihood and gradients and then use scipy\'s minimize function for

3条回答
  •  醉酒成梦
    2021-01-05 09:14

    I had similar problem. I had an iteration, and sometimes execution took so long it timed out. Increasing spark.executor.heartbeatInterval seemed to solve the problem. I increased it to 3600s to ensure I don't run into timeouts again and everything is working fine since then.

    From: http://spark.apache.org/docs/latest/configuration.html :

    spark.executor.heartbeatInterval 10s Interval between each executor's heartbeats to the driver. Heartbeats let the driver know that the executor is still alive and update it with metrics for in-progress tasks.

提交回复
热议问题