问题 I wish to connect to a remote cluster and execute a Spark process. So, from what I have read, this is specified in the SparkConf. val conf = new SparkConf() .setAppName("MyAppName") .setMaster("spark://my_ip:7077") Where my_ip is the IP address of my cluster. Unfortunately, I get connection refused. So, I am guessing some credentials must be added to connect correctly. How would I specify the credentials? It seems it would be done with .set(key, value), but have no leads on this. 回答1: There