【发布时间】:2016-06-23 23:52:30
【问题描述】:
我是 Spark 的新手。刚刚在我的笔记本电脑(Ubuntu)中安装了 1.6.0 版,并继续使用 phyton 的外壳(pyspark)进行示例。但是,我无法弄清楚这个错误告诉我什么。你能帮忙吗?任何帮助表示赞赏。
>>> lines = sc.textFile("spark-1.6.0/README.md")
>>> lines.count()
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/ricky/spark-1.6.0/python/pyspark/rdd.py", line 1004, in count
return self.mapPartitions(lambda i: [sum(1 for _ in i)]).sum()
File "/home/ricky/spark-1.6.0/python/pyspark/rdd.py", line 995, in sum
return self.mapPartitions(lambda x: [sum(x)]).fold(0, operator.add)
File "/home/ricky/spark-1.6.0/python/pyspark/rdd.py", line 869, in fold
vals = self.mapPartitions(func).collect()
File "/home/ricky/spark-1.6.0/python/pyspark/rdd.py", line 771, in collect
port = self.ctx._jvm.PythonRDD.collectAndServe(self._jrdd.rdd())
File "/home/ricky/spark-1.6.0/python/lib/py4j-0.9- src.zip/py4j/java_gateway.py", line 813, in __call__
File "/home/ricky/spark-1.6.0/python/pyspark/sql/utils.py", line 45, in deco
return f(*a, **kw)
File "/home/ricky/spark-1.6.0/python/lib/py4j-0.9- src.zip/py4j/protocol.py", line 308, **in get_return_value**
py4j.protocol.Py4JJavaError: An error occurred while calling z:org.apache.spark.api.python.PythonRDD.collectAndServe.
: java.net.BindException: Cannot assign requested address
at java.net.PlainSocketImpl.socketBind(Native Method)
at java.net.AbstractPlainSocketImpl.bind(AbstractPlainSocketImpl.java:387)
at java.net.ServerSocket.bind(ServerSocket.java:375)
at java.net.ServerSocket.<init>(ServerSocket.java:237)
at org.apache.spark.api.python.PythonRDD$.serveIterator(PythonRDD.scala:637)
at org.apache.spark.api.python.PythonRDD$.collectAndServe(PythonRDD.scala:405)
at org.apache.spark.api.python.PythonRDD.collectAndServe(PythonRDD.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java :43)
at java.lang.reflect.Method.invoke(Method.java:498)
at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:231)
at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:381)
at py4j.Gateway.invoke(Gateway.java:259)
at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133)
at py4j.commands.CallCommand.execute(CallCommand.java:79)
at py4j.GatewayConnection.run(GatewayConnection.java:209)
at java.lang.Thread.run(Thread.java:745)
【问题讨论】:
标签: apache-spark pyspark