gpt4 book ai didi

mysql - CoarseGrainedExecutorBackend ERROR Spark

转载 作者:行者123 更新时间:2023-11-30 21:58:42 27 4
gpt4 key购买 nike

我使用 Spark 2.1.1 和 Scala 2.10 以及两个节点的 Spark Standalone。

我想执行从 MySQL 表中读取数据集并将其写入 Cassandra 表的 Spark 代码。

代码:

object RdmsToCassandra extends Serializable {

def main(args: Array[String]) {
val conf = new SparkConf().set("spark.cassandra.connection.host", "192.168.22.231")
val sc = new SparkContext("spark://192.168.22.231:7077", "MigrateMySQLToCassandra", conf)
val mysqlJdbcString: String = s"jdbc:mysql://192.168.22.238/customer_events?user=root&password=qweqwe"
Class.forName("com.mysql.jdbc.Driver").newInstance

CassandraConnector(conf).withSessionDo { session =>
session.execute("CREATE KEYSPACE IF NOT EXISTS test WITH replication = {'class': 'SimpleStrategy', 'replication_factor': 2 }")
session.execute("CREATE TABLE IF NOT EXISTS test.store( store_name text PRIMARY KEY, location text, store_type text)" )
}
val highestId: Long = 2
val startingId: Long = 0
val numberOfPartitions = 1;

val customerEvents = new JdbcRDD(sc, () => { DriverManager.getConnection(mysqlJdbcString)},
"select * from store limit ?, ?",startingId, highestId, numberOfPartitions,
(r: ResultSet) => {
(r.getString("store_name"),
r.getString("location"),
r.getString("store_type")
)
}
)


customerEvents.saveToCassandra("test", "store1",
SomeColumns("store_name"))

}

}

我使用命令提交申请:

spark-submit --master spark://192.168.22.231:6066 \
--class "RdmsToCassandra" \
rdbmstocassandra_2.10-1.0.jar

在执行过程中出现以下错误 _

Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties 17/05/26 15:45:51 INFO CoarseGrainedExecutorBackend: Started daemon with process name: 10619@totalprices-db-server-02 17/05/26 15:45:51 INFO SignalUtils: Registered signal handler for TERM 17/05/26 15:45:51 INFO SignalUtils: Registered signal handler for HUP 17/05/26 15:45:51 INFO SignalUtils: Registered signal handler for INT 17/05/26 15:45:51 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 17/05/26 15:45:51 INFO SecurityManager: Changing view acls to: root 17/05/26 15:45:51 INFO SecurityManager: Changing modify acls to: root 17/05/26 15:45:51 INFO SecurityManager: Changing view acls groups to: 17/05/26 15:45:51 INFO SecurityManager: Changing modify acls groups to: 17/05/26 15:45:51 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); groups with view permissions: Set(); users with modify permissions: Set(root); groups with modify permissions: Set() 17/05/26 15:45:52 INFO TransportClientFactory: Successfully created connection to /192.168.22.231:36249 after 60 ms (0 ms spent in bootstraps) 17/05/26 15:45:52 INFO SecurityManager: Changing view acls to: root 17/05/26 15:45:52 INFO SecurityManager: Changing modify acls to: root 17/05/26 15:45:52 INFO SecurityManager: Changing view acls groups to: 17/05/26 15:45:52 INFO SecurityManager: Changing modify acls groups to: 17/05/26 15:45:52 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(root); groups with view permissions: Set(); users with modify permissions: Set(root); groups with modify permissions: Set() 17/05/26 15:45:52 INFO TransportClientFactory: Successfully created connection to /192.168.22.231:36249 after 1 ms (0 ms spent in bootstraps) 17/05/26 15:45:52 WARN Utils: Your hostname, totalprices-db-server-02 resolves to a loopback address: 127.0.0.1; using 221.243.36.126 instead (on interface em1) 17/05/26 15:45:52 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to another address 17/05/26 15:45:52 INFO DiskBlockManager: Created local directory at /tmp/spark-73513e64-f52c-48c5-bf9e-bbc45caec12d/executor-fd21cdc3-673e-4b9e-9bd2-6cef1e3da140/blockmgr-157ab1c4-ca1e-481b-9357-b5855ee6beef 17/05/26 15:45:52 INFO MemoryStore: MemoryStore started with capacity 2004.6 MB 17/05/26 15:45:52 INFO CoarseGrainedExecutorBackend: Connecting to driver: spark://CoarseGrainedScheduler@192.168.22.231:36249 17/05/26 15:45:52 INFO WorkerWatcher: Connecting to worker spark://Worker@221.243.36.126:40561 17/05/26 15:45:52 INFO TransportClientFactory: Successfully created connection to /221.243.36.126:40561 after 1 ms (0 ms spent in bootstraps) 17/05/26 15:45:52 INFO WorkerWatcher: Successfully connected to spark://Worker@221.243.36.126:40561 17/05/26 15:45:52 INFO CoarseGrainedExecutorBackend: Successfully registered with driver 17/05/26 15:45:52 INFO Executor: Starting executor ID 0 on host 221.243.36.126 17/05/26 15:45:52 INFO Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 43114. 17/05/26 15:45:52 INFO NettyBlockTransferService: Server created on 221.243.36.126:43114 17/05/26 15:45:52 INFO BlockManager: Using org.apache.spark.storage.RandomBlockReplicationPolicy for block replication policy 17/05/26 15:45:52 INFO BlockManagerMaster: Registering BlockManager BlockManagerId(0, 221.243.36.126, 43114, None) 17/05/26 15:45:52 INFO BlockManagerMaster: Registered BlockManager BlockManagerId(0, 221.243.36.126, 43114, None) 17/05/26 15:45:52 INFO BlockManager: Initialized BlockManager: BlockManagerId(0, 221.243.36.126, 43114, None) 17/05/26 15:46:00 INFO CoarseGrainedExecutorBackend: Driver commanded a shutdown 17/05/26 15:46:00 ERROR CoarseGrainedExecutorBackend: RECEIVED SIGNAL TERM tdown

任何人都可以提出这里可能存在的问题。

最佳答案

我知道它有点旧,但可能对路过的其他人有帮助:

我遇到了类似的问题,对 NodeManager-Logs 的调查让我找到了根本原因。您可以通过 Web 界面访问它们

nodeManagerAdress:PORT/logs

PORTyarn.nodemanager.webapp.address 下的 yarn-site.xml 中指定。 (默认:8042)

我的调查工作流程:

  1. 收集日志( yarn 日志...命令)
  2. 识别节点和容器(在这些日志中)发出错误
  3. 错误的时间戳搜索 NodeManager 日志以查找根本原因

顺便说一句:您可以访问影响同一端口节点的所有配置的聚合集合 (xml):

 nodeManagerAdress:PORT/conf

关于mysql - CoarseGrainedExecutorBackend ERROR Spark ,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/44195034/

27 4 0
Copyright 2021 - 2024 cfsdn All Rights Reserved 蜀ICP备2022000587号
广告合作:1813099741@qq.com 6ren.com