Azure Data Explorer ingests duplicated records from Event Hub

J.P. Zhang 25 Reputation points Microsoft Employee
2025-03-14T17:59:18.19+00:00

We have data ingestions set up between an event hub and Azure Data Explorer. There are abnormally high duplicated records in Azure Data Explorer database tables. I understand that the ingestion delivery is at least once. having dups are sort of expected. But we are seeing very high dups at a rate of 50 dups out of 160 records. What could be the issue here? What configuration changes can be made to mitigate this?

Azure Data Explorer
Azure Data Explorer
An Azure data analytics service for real-time analysis on large volumes of data streaming from sources including applications, websites, and internet of things devices.
561 questions
{count} votes

Accepted answer
  1. Sander van de Velde | MVP 36,146 Reputation points MVP
    2025-03-17T18:26:05.6366667+00:00

    Hello @J.P. Zhang,

    welcome to this moderated Azure community forum.

    The data ingestions via Data connections set up between an event hub and Azure Data Explorer showed an abnormally high rate of duplicated records in Azure Data Explorer database tables (50 dups out of 160 records).

    The solution chosen here was deleting all data connections and recreated them again.

    After that, the test runs no longer show duplicates.

    Azure Eventhubs have an option to disable sending:

    User's image

    If the duration of disabling sending data is less than the retention time set of the eventhub, not data is lost, only delayed.

    This does not touch any schema or data already ingested in the Azure Data Explorer.


    If the response helped, do "Accept Answer". If it doesn't work, please let us know the progress. All community members with similar issues will benefit by doing so. Your contribution is highly appreciated.

    0 comments No comments

0 additional answers

Sort by: Most helpful

Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.