Spark + s3 - error - java.lang.ClassNotFoundException: Class org.apache.hadoop.fs.s3a.S3AFileSystem not found

前端 未结 4 1171
抹茶落季
抹茶落季 2020-12-20 14:35

I have a spark ec2 cluster where I am submitting a pyspark program from a Zeppelin notebook. I have loaded the hadoop-aws-2.7.3.jar and aws-java-sdk-1.11.179.jar and place

4条回答
  •  北海茫月
    2020-12-20 14:42

    Add the following to this file hadoop/etc/hadoop/core-site.xml

    
      fs.s3.awsAccessKeyId
      ***
    
    
      fs.s3.awsSecretAccessKey
      ***
    
    

    Inside the Hadoop installation directory, find aws jars, for MAC installation directory is /usr/local/Cellar/hadoop/

    find . -type f -name "*aws*"
    
    sudo cp hadoop/share/hadoop/tools/lib/aws-java-sdk-1.7.4.jar hadoop/share/hadoop/common/lib/
    sudo cp hadoop/share/hadoop/tools/lib/hadoop-aws-2.7.5.jar hadoop/share/hadoop/common/lib/
    

    Credit

提交回复
热议问题