Can a PySpark Kernel(JupyterHub) run in yarn-client mode?
问题 My Current Setup: Spark EC2 Cluster with HDFS and YARN JuputerHub(0.7.0) PySpark Kernel with python27 The very simple code that I am using for this question: rdd = sc.parallelize([1, 2]) rdd.collect() The PySpark kernel that works as expected in Spark standalone has the following environment variable in the kernel json file: "PYSPARK_SUBMIT_ARGS": "--master spark://<spark_master>:7077 pyspark-shell" However, when I try to run in yarn-client mode it is getting stuck forever, while the log