gpt4 book ai didi

azure - Databricks 连接 : can't connect to remote cluster on azure, 命令: 'databricks-connect test' 停止

转载 作者:行者123 更新时间:2023-12-03 21:02:41 25 4
gpt4 key购买 nike

我尝试设置 Databricks Connect 以便能够与已在 Azure 上的工作区上运行的远程 Databricks 集群配合使用。当我尝试运行命令:“databricks-connect test”时,它永远不会结束。

我关注官方documentation .

我已经安装了最新的 Anaconda 版本 3.7。我已经创建了本地环境: conda create --name dbconnect python=3.5

我在 5.1 版中安装了“databricks-connect”,它与 Azure Databricks 上的集群配置相匹配。

    pip install -U databricks-connect==5.1.*

我已经按如下方式设置了“databricks-connect 配置”:

    (base) C:\>databricks-connect configure
The current configuration is:
* Databricks Host: ******.azuredatabricks.net
* Databricks Token: ************************************
* Cluster ID: ****-******-*******
* Org ID: ****************
* Port: 8787

经过上述步骤后,我尝试运行 databricks 连接的“测试”命令:

    databricks-connect test

因此,在有关 MetricsSystem 的警告后,程序开始并停止,如下所示:

    (dbconnect) C:\>databricks-connect test
* PySpark is installed at c:\users\miltad\appdata\local\continuum\anaconda3\envs\dbconnect\lib\site-packages\pyspark
* Checking java version
java version "1.8.0_181"
Java(TM) SE Runtime Environment (build 1.8.0_181-b13)
Java HotSpot(TM) 64-Bit Server VM (build 25.181-b13, mixed mode)
* Testing scala command
19/05/31 08:14:26 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
19/05/31 08:14:34 WARN MetricsSystem: Using default name SparkStatusTracker for source because neither spark.metrics.namespace nor spark.app.id is set.

我希望该流程应该像官方 documentation 中那样进入下一步:

    * Testing scala command
18/12/10 16:38:44 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
18/12/10 16:38:50 WARN MetricsSystem: Using default name SparkStatusTracker for source because neither spark.metrics.namespace nor spark.app.id is set.
18/12/10 16:39:53 WARN SparkServiceRPCClient: Now tracking server state for 5abb7c7e-df8e-4290-947c-c9a38601024e, invalidating prev state
18/12/10 16:39:59 WARN SparkServiceRPCClient: Syncing 129 files (176036 bytes) took 3003 ms
Welcome to
____ __
/ __/__ ___ _____/ /__
_\ \/ _ \/ _ `/ __/ '_/
/___/ .__/\_,_/_/ /_/\_\ version 2.4.0-SNAPSHOT
/_/

Using Scala version 2.11.12 (Java HotSpot(TM) 64-Bit Server VM, Java 1.8.0_152)
Type in expressions to have them evaluated.
Type :help for more information.

因此,我的进程在“WARN MetricsSystem:使用默认名称 SparkStatusTracker”之后停止。

我做错了什么?我应该配置更多东西吗?

最佳答案

运行时 5.3 或更低版本似乎不正式支持此功能。如果更新运行时有限制,我将确保 Sparkconf 设置如下:spark.databricks.service.server.enabled true然而,对于较旧的运行时,事情仍然可能不稳定。我建议使用运行时 5.5 或 6.1 或更高版本执行此操作。

关于azure - Databricks 连接 : can't connect to remote cluster on azure, 命令: 'databricks-connect test' 停止,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/56389816/

25 4 0
Copyright 2021 - 2024 cfsdn All Rights Reserved 蜀ICP备2022000587号
广告合作:1813099741@qq.com 6ren.com