My Spark's Worker cannot connect Master.Something wrong with Akka?

前端 未结 7 834
萌比男神i
萌比男神i 2020-12-15 23:46

I want to install Spark Standlone mode to a Cluster with my two virtual machines.
With the version of spark-0.9.1-bin-hadoop1, I execute spark-shell successfully in each

相关标签:
7条回答
  • 2020-12-16 00:14

    There's a lot of answers and possible solutions, and this question is a bit old, but in the interest of completeness, there is a known Spark bug about hostnames resolving to IP addresses. I am not presenting this as the complete answer in all cases, but I suggest trying with a baseline of just using all IPs, and only use the single config SPARK_MASTER_IP. With just those two practices I get my clusters to work and all the other configs, or using hostnames, just seems to muck things up.

    So in your spark-env.sh get rid of SPARK_WORKER_IP and change SPARK_MASTER_IP to an IP address, not a hostname.

    I have treated this more at length in this answer.

    For more completeness here's part of that answer:

    Can you ping the box where the Spark master is running? Can you ping the worker from the master? More importantly, can you password-less ssh to the worker from the master box? Per 1.5.2 docs you need to be able to do that with a private key AND have the worker entered in the conf/slaves file. I copied the relevant paragraph at the end.

    You can get a situation where the worker can contact the master but the master can't get back to the worker so it looks like no connection is being made. Check both directions.

    I think the slaves file on the master node, and the password-less ssh can lead to similar errors to what you are seeing.

    Per the answer I crosslinked, there's also an old bug but it's not clear how that bug was resolved.

    0 讨论(0)
提交回复
热议问题