“Container killed by YARN for exceeding memory limits. 10.4 GB of 10.4 GB physical memory used” on an EMR cluster with 75GB of memory

前端 未结 5 633
梦谈多话
梦谈多话 2020-12-04 05:19

I\'m running a 5 node Spark cluster on AWS EMR each sized m3.xlarge (1 master 4 slaves). I successfully ran through a 146Mb bzip2 compressed CSV file and ended up with a per

5条回答
  •  情话喂你
    2020-12-04 05:56

    If you're not using spark-submit, and you're looking for another way to specify the yarn.nodemanager.vmem-check-enabled parameter mentioned by Duff, here are 2 other ways:

    Method 2

    If you're using a JSON Configuration file (that you pass to the AWS CLI or to your boto3 script), you'll have to add the following configuration:

    [{
    "Classification": "yarn-site", 
      "Properties": {
        "yarn.nodemanager.vmem-check-enabled": "false"
       }
    }]
    

    Method 3

    If you use the EMR console, add the following configuration:

    classification=yarn-site,properties=[yarn.nodemanager.vmem-check-enabled=false]
    

提交回复
热议问题