google-cloud-dataproc

Can't create a Python 3 notebook in jupyter notebook

六眼飞鱼酱① 提交于 2020-12-10 06:59:40
问题 I'm following this tutorial and I'm stuck when I want to create a new Jupyter Notebook (Python 3). The cluster is created using this command: gcloud beta dataproc clusters create ${CLUSTER_NAME} \ --region=${REGION} \ --image-version=1.4 \ --master-machine-type=n1-standard-4 \ --worker-machine-type=n1-standard-4 \ --bucket=${BUCKET_NAME} \ --optional-components=ANACONDA,JUPYTER \ --enable-component-gateway When I accessing the JupyterLab and try to create a new notebook I can see: and then

Can I run dataproc jobs in cluster mode

泄露秘密 提交于 2020-07-19 06:45:27
问题 Just starting to get familiar with GCP dataproc. I've noticed when I use gcloud dataproc jobs submit pyspark that jobs are submitted with spark.submit.deployMode=client . Is spark.submit.deployMode=cluster an option for us? 回答1: Yes, you can, by specifying --properties spark.submit.deployMode=cluster . Just note that driver output will be in yarn userlogs (you can access them in Stackdriver Logging from the Console). We run in client mode by default to stream driver output to you. 来源: https:/

Error when running python map reduce job using Hadoop streaming in Google Cloud Dataproc environment

强颜欢笑 提交于 2020-07-05 04:55:34
问题 I want to run python map reduce job in Google Cloud Dataproc using hadoop streaming method. My map reduce python script, input file and job result output are located in Google Cloud Storage. I tried to run this command hadoop jar /usr/lib/hadoop-mapreduce/hadoop-streaming.jar -file gs://bucket-name/intro_to_mapreduce/mapper_prod_cat.py -mapper gs://bucket-name/intro_to_mapreduce/mapper_prod_cat.py -file gs://bucket-name/intro_to_mapreduce/reducer_prod_cat.py -reducer gs://bucket-name/intro_to