gpt4 book ai didi

apache-spark - 类型错误 : 'JavaPackage' object is not callable

转载 作者:行者123 更新时间:2023-12-04 04:40:21 25 4
gpt4 key购买 nike

当我编写 spark sql API hiveContext.sql()

from pyspark import SparkConf,SparkContext
from pyspark.sql import SQLContext,HiveContext

conf = SparkConf().setAppName("spark_sql")

sc = SparkContext(conf = conf)
hc = HiveContext(sc)

#rdd = sc.textFile("test.txt")
sqlContext = SQLContext(sc)
res = hc.sql("use teg_uee_app")
#for each in res.collect():
# print(each[0])
sc.stop()

我收到以下错误:
enFile "spark_sql.py", line 23, in <module>
res = hc.sql("use teg_uee_app")
File "/spark/python/pyspark/sql/context.py", line 580, in sql
return DataFrame(self._ssql_ctx.sql(sqlQuery), self)
File "/spark/python/pyspark/sql/context.py", line 683, in _ssql_ctx
self._scala_HiveContext = self._get_hive_ctx()
File "/spark/python/pyspark/sql/context.py", line 692, in _get_hive_ctx
return self._jvm.HiveContext(self._jsc.sc())
TypeError: 'JavaPackage' object is not callable

我如何添加 SPARK_CLASSPATH 或 SparkContext.addFile?我不知道。

最佳答案

也许这会对您有所帮助:使用 HiveContext 时,我必须向 spark-submit 参数添加三个 jar:

spark-submit --jars /usr/lib/spark/lib/datanucleus-api-jdo-3.2.6.jar,/usr/lib/spark/lib/datanucleus-core-3.2.10.jar,/usr/lib/spark/lib/datanucleus-rdbms-3.2.9.jar ...

当然,路径和版本取决于您的集群设置。

关于apache-spark - 类型错误 : 'JavaPackage' object is not callable,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/38204806/

25 4 0
Copyright 2021 - 2024 cfsdn All Rights Reserved 蜀ICP备2022000587号
广告合作:1813099741@qq.com 6ren.com