Spark java.lang.OutOfMemoryError: Java heap space

后端 未结 12 2168
半阙折子戏
半阙折子戏 2020-11-22 13:55

My cluster: 1 master, 11 slaves, each node has 6 GB memory.

My settings:

spark.executor.memory=4g, Dspark.akka.frameSize=512

12条回答
  •  旧巷少年郎
    2020-11-22 14:13

    Have a look at the start up scripts a Java heap size is set there, it looks like you're not setting this before running Spark worker.

    # Set SPARK_MEM if it isn't already set since we also use it for this process
    SPARK_MEM=${SPARK_MEM:-512m}
    export SPARK_MEM
    
    # Set JAVA_OPTS to be able to load native libraries and to set heap size
    JAVA_OPTS="$OUR_JAVA_OPTS"
    JAVA_OPTS="$JAVA_OPTS -Djava.library.path=$SPARK_LIBRARY_PATH"
    JAVA_OPTS="$JAVA_OPTS -Xms$SPARK_MEM -Xmx$SPARK_MEM"
    

    You can find the documentation to deploy scripts here.

提交回复
热议问题