Wir müssen einige große Dateien in Azure Data Lake Speicher von verschachtelten JSON in CSV konvertieren konvertieren. Da die Python-Module pandas, numpy in Azure data lake analytics neben den Standard-Modulen unterstützt werden, ist es meines Erachtens möglich, dies mit Python zu erreichen. Hat jemand den Python-Code, um dies zu erreichen?U-SQL mit Python zum Konvertieren von JSON zu CSV in Azure Data Lake speichern
Quelle Format:
{ "Loc": "TDM", "Thema": "Lage", "LocMac": "Lage/fe: 7a: xx: xx: xx: xx "," seq ":" 296083773 "," Zeitstempel ": 1488986751," op ":" OP_UPDATE "," topicSeq ":" 46478211 "," sourceId ":" AFBWmHs "," Standort ": {" staEthMac ": { "addr": "/ xxxxx"}, "staLocationX": 1643.8915, "staLocationY": 571.04205, "errorLevel": 1076, "assoziiert": 0, "campusId": "n5THo6IINuOSVZ/cTidNVA ==", "buildingId": "7hY/xx ==", "floorId": "xxxxxxxxxx + BYoo0A ==", "hashedStaEthMac": "xxxx/pMVyK4Gu9qG6w =", "locAlgorithm": "ALGORITHM_ESTIMATION", "Einheit": "FEET"}, "EventProcessedUtcTime ":" 2017-03-08T15: 35: 02.3847947Z "," PartitionId ": 3," EventEnqueuedUtcTime ":" 2017-03-08T15: 35: 03.7510000Z "," IoTHub ": {" Mess ageId ": null," CorrelationId ": null," ConnectionDeviceId ":" xxxxx "," ConnectionDeviceGenerationId ":" 636243184116591838 "," EnqueuedTime ":" 0001-01-01T00: 00: 00.0000000 "," StreamId ": null}}
Erwartete Ausgabe
TDM, Lage, Lage/80: 7a: bf: d4: d6: 50,974851970,1490004475, OP_UPDATE, 151.002.334, xxxxxxx, ghq/1NZQ, 977,7259, 638.8827,490,1, n5THo6IINuOSVZ/cTidNVA ==, 7HY/jVh9NRqqxF6gbqT7Jw ==, LV/ZiQRQMS2wwKiKTvYNBQ ==, H5rrAD/jg1Fnkmo1Zmquau/Qn1U =, ALGORITHM_ESTIMATION, Füße