Ingesting json events with various schemas from Blob Storage to Azure Data Explorer

Francois 26 Reputation points
2020-10-12T15:45:39.94+00:00

Hi,

We're a game company using PlayFab for our backend. Play sessions events are created by PlayFab and are automatically sent to a blob storage in Azure. I created an Azure Data Explorer Database which ingests these events whenever the blob storage receives them. To ingest the data, I've created a database mapping and a data connection as shown in this tutorial: https://video2.skills-academy.com/en-us/azure/data-explorer/ingest-data-event-grid

Ingestion is globally working properly, but I realized some events were not properly ingested. It turns out some PlayFab events are using a different JSON format, as indicated by a SchemaVersion field. For these events, the mapping should be different (columns should correspond to a different name in the json file). I would therefore need to change the mapping based on that field. Is that something that can be done using the current ingestion method we're using? I've seen that other Azure services seem to support this type of conditional mapping, but nothing for Kusto in Azure Data Explorer. Ultimately we need the data to be in Azure Data Explorer to analyze events through queries.

Thanks a lot for your help

Azure Data Explorer
Azure Data Explorer
An Azure data analytics service for real-time analysis on large volumes of data streaming from sources including applications, websites, and internet of things devices.
502 questions
{count} votes

Accepted answer
  1. HarithaMaddi-MSFT 10,136 Reputation points
    2020-10-15T07:00:57.093+00:00

    Hi @Francois ,

    Thanks for the detailed response.

    • Since there is no information on specific pattern of file names, prefix option is ruled out
    • Product team confirmed that multiple potential fields is not feasible today. They suggested the feasible thing in this scenario can be to ingest entire records as “dynamic” field in one table with low retention and set up an update policy (or several) to split the data as required.
    • Yes, Power Automate is also an option but from ETL aspect, Azure data factory is more powerful than Power Automate and is reasonable in cost when I looked at pricing of both the tools. I understand the challenge of using another service, please suggest if above alternative using update policy will be helpful.
    1 person found this answer helpful.

0 additional answers

Sort by: Most helpful