gpt4 book ai didi

json - 如何在Scala Spark 2.0中解析json文件,如何将数据插入Hive表中?

转载 作者:行者123 更新时间:2023-12-02 22:09:11 25 4
gpt4 key购买 nike

我想在spark 2.0(scala)中解析json文件。接下来,我想将数据保存在Hive表中。
如何使用Scala解析json文件?
json文件示例)metadata.json:

  {
"syslog": {
"month": "Sep",
"day": "26",
"time": "23:03:44",
"host": "cdpcapital.onmicrosoft.com"
},
"prefix": {
"cef_version": "CEF:0",
"device_vendor": "Microsoft",
"device_product": "SharePoint Online",
},
"extensions": {
"eventId": "7808891",
"msg": "ManagedSyncClientAllowed",
"art": "1506467022378",
"cat": "SharePoint",
"act": "ManagedSyncClientAllowed",
"rt": "1506466717000",
"requestClientApplication": "Microsoft SkyDriveSync",
"cs1": "0bdbe027-8f50-4ec3-843f-e27c41a63957",
"cs1Label": "Organization ID",
"cs2Label": "Modified Properties",
"ahost": "cdpdiclog101.cgimss.com",
"agentZoneURI": "/All Zones",
"amac": "F0-1F-AF-DA-8F-1B",
"av": "7.6.0.8009.0",
}
},

谢谢

最佳答案

您可以使用类似:

val jsonDf = sparkSession
.read
//.option("wholeFile", true) if its not a Single Line JSON
.json("resources/json/metadata.json")

jsonDf.printSchema()

jsonDf.registerTempTable("metadata")

有关此 https://spark.apache.org/docs/latest/sql-programming-guide.html#hive-tables的更多详细信息

关于json - 如何在Scala Spark 2.0中解析json文件,如何将数据插入Hive表中?,我们在Stack Overflow上找到一个类似的问题: https://stackoverflow.com/questions/47339601/

25 4 0
Copyright 2021 - 2024 cfsdn All Rights Reserved 蜀ICP备2022000587号
广告合作:1813099741@qq.com 6ren.com