【问题标题】:Worker: Failed to connect to master on spark apache工作人员:无法连接到 spark apache 上的 master
【发布时间】:2017-09-26 11:03:15
【问题描述】:

我正在尝试在 Windows 上使用独立集群管理器部署 Spark 应用程序,因此在主端(IP:192...186)我运行:

\bin>spark-class org.apache.spark.deploy.master.Master

在从机端(IP:192...75)我跑了:

\bin>spark-class org.apache.spark.deploy.worker.Worker spark://192.*.*.186:7077

我有一个问题,客户端无法连接到主节点,这是我得到的错误:

17/09/26 12:00:51 WARN Worker: Failed to connect to master 192.*.*.186:7077
org.apache.spark.SparkException: Exception thrown in awaitResult:
    at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:205)
    at org.apache.spark.rpc.RpcTimeout.awaitResult(RpcTimeout.scala:75)
    at org.apache.spark.rpc.RpcEnv.setupEndpointRefByURI(RpcEnv.scala:100)
    at org.apache.spark.rpc.RpcEnv.setupEndpointRef(RpcEnv.scala:108)
    at org.apache.spark.deploy.worker.Worker$$anonfun$org$apache$spark$deploy$worker$Worker$$tryRegisterAllMasters$1$$anon$1.run(Worker.scala:241)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.run(FutureTask.java:266)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.io.IOException: Failed to connect to /192.*.*.186:7077
    at org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:232)
    at org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:182)
    at org.apache.spark.rpc.netty.NettyRpcEnv.createClient(NettyRpcEnv.scala:197)
    at org.apache.spark.rpc.netty.Outbox$$anon$1.call(Outbox.scala:194)
    at org.apache.spark.rpc.netty.Outbox$$anon$1.call(Outbox.scala:190)
    ... 4 more
Caused by: io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection refused: no further information: /192.*.*.186:7077
    at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
    at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
    at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(NioSocketChannel.java:257)
    at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:291)
    at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:631)
    at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:566)
    at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:480)
    at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:442)
    at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131)
    at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
    ... 1 more

我该如何解决这个问题?

【问题讨论】:

  • 测试两台机器之间的连接?

标签: apache-spark worker master apache-spark-standalone


【解决方案1】:
bin>spark-class org.apache.spark.deploy.worker.Worker spark://192.*.*.186:7077

上面命令中提到的master-ip是否正确?

您写道主 ip 是 192...168 但在从属启动命令中您正在尝试使用 192...186 ,错字:6 和 8 倒置

【讨论】:

  • 好的,那么你可以尝试使用以下命令从从机远程登录到主机。 > 远程登录主 IP 7077
  • 另外,在文件中添加以下行:SPARK_HOME/conf/spark-env.sh SPARK_MASTER_IP=192...186 然后重新启动 spark master
  • ^ 在 master 中:在文件 SPARK_HOME/conf/spark-env.sh SPARK_MASTER_IP=192...186
  • 我已经尝试过了,但是我收到了一个奇怪的行为:在主节点 Spark Web UI 中,部署主节点的 url 不是机器 ip 地址:URL:spark:/ /169.254.191.139:7077 REST URL:spark://169.254.191.139:6066(集群模式)Alive Workers:0
猜你喜欢
  • 2020-11-08
  • 1970-01-01
  • 2018-02-17
  • 1970-01-01
  • 1970-01-01
  • 2018-04-10
  • 2021-09-10
  • 1970-01-01
  • 2023-03-11
相关资源
最近更新 更多