Why does join fail with “java.util.concurrent.TimeoutException: Futures timed out after [300 seconds]”?

前端 未结 4 1443
Happy的楠姐
Happy的楠姐 2020-11-30 19:33

I am using Spark 1.5.

I have two dataframes of the form:

scala> libriFirstTable50Plus3DF
res1: org.apache.spark.sql.DataFrame = [basket_id: string         


        
4条回答
  •  青春惊慌失措
    2020-11-30 19:59

    In addition to increasing spark.sql.broadcastTimeout or persist() both DataFrames,

    You may try:

    1.disable broadcast by setting spark.sql.autoBroadcastJoinThreshold to -1

    2.increase the spark driver memory by setting spark.driver.memory to a higher value.

提交回复
热议问题