Unable to execute spark job using SparkSubmitOperator

后端 未结 1 851
自闭症患者
自闭症患者 2021-02-06 03:28

I am able to run Spark job using BashOperator but I want to use SparkSubmitOperator for it using Spark standalone mod

1条回答
  •  一整个雨季
    2021-02-06 03:54

    You can either create a new connection using the Airflow Web UI or change the spark-default connection.

    Master can be local, yarn, spark://HOST:PORT, mesos://HOST:PORT and k8s://https://:.

    You can also supply the following commands in the extras:

    {"queue": "root.default", "deploy_mode": "cluster", "spark_home": "", "spark_binary": "spark-submit", "namespace": "default"}
    

    Either the "spark-submit" binary should be in the PATH or the spark-home is set in the extra on the connection.

    0 讨论(0)
提交回复
热议问题