【发布时间】:2018-07-12 13:13:30
【问题描述】:
spark-submit --packages com.databricks:spark-redshift_2.11:2.0.1 --jars /usr/share/aws/redshift/jdbc/RedshiftJDBC4.jar /home/hadoop/test.py
如何以 Apache livy 格式指定上述 (pyspark) spark-submit 命令?
我尝试了以下方法:
curl -X POST --data '{"file": "/home/hadoop/test.py", "conf":
{"com.databricks": "spark-redshift_2.11:2.0.1"}, \
"queue": "my_queue", "name": "Livy Example", "jars" :
"/usr/share/aws/redshift/jdbc/RedshiftJDBC4.jar"}', \
-H "Content-Type: application/json" localhost:8998/batches
Refered the following livy article spark livy rest api
同样出现以下错误:
"Unexpected character ('“' (code 8220 / 0x201c)): was expecting double-quote to start field name\n at [Source: (org.eclipse.jetty.server.HttpInputOverHTTP); line: 1, column: 37]
【问题讨论】:
-
您是否剪切和粘贴错误,或者您的数据中有那些智能引号? (参见
conf和com.databricks...) -
让我检查一下...
-
还是同样的错误。
标签: apache-spark hadoop pyspark amazon-emr