问题
Just starting to get familiar with GCP dataproc. I've noticed when I use gcloud dataproc jobs submit pyspark
that jobs are submitted with spark.submit.deployMode=client
. Is spark.submit.deployMode=cluster
an option for us?
回答1:
Yes, you can, by specifying --properties spark.submit.deployMode=cluster
. Just note that driver output will be in yarn userlogs (you can access them in Stackdriver Logging from the Console). We run in client mode by default to stream driver output to you.
来源:https://stackoverflow.com/questions/49678757/can-i-run-dataproc-jobs-in-cluster-mode