【发布时间】:2016-05-25 05:50:27
【问题描述】:
我尝试使用以下命令运行 python spark-shell:
bin/pyspark --packages datastax:spark-cassandra-connector:1.5.0-RC1-s_2.11,org.apache.spark:spark-streaming-kafka_2.10:1.6.0
以下命令的输出表明它能够找到 spark-cassandra-connector 包:
resolving dependencies :: org.apache.spark#spark-submit-parent;1.0
confs: [default]
found datastax#spark-cassandra-connector;1.5.0-RC1-s_2.11 in spark-packages
found org.apache.cassandra#cassandra-clientutil;2.2.2 in central
found com.datastax.cassandra#cassandra-driver-core;3.0.0-rc1 in central
found io.netty#netty-handler;4.0.33.Final in central
found io.netty#netty-buffer;4.0.33.Final in central
found io.netty#netty-common;4.0.33.Final in central
但是当我尝试使用以下任何命令导入包时,我得到导入错误:
from com.datastax import *
from com.datastax.spark.connector import *
输出:
ImportError: No module named com.datastax
ImportError: No module named com.datastax.spark.connector
谁能建议这里出了什么问题?
【问题讨论】:
-
你好,你是怎么解决这个问题的?请说我
标签: python pyspark spark-streaming datastax spark-cassandra-connector