Forum Discussion

bharath_janjanam's avatar
bharath_janjanam
Copper Contributor
Jul 20, 2022

Data Sink From Dataverse to Azure SQL database using azure data factory is taking really long time

Hi Everyone,

 

Actually I'm new to the azure data factory, we are facing 2 problem while sinking data verse to azure SQL database for reporting purpose.

 

Problem 1:

We have created a azure synapse link to push the data to the azure data lake from there we are pushing data to Azure SQL DB using Azure Data Factory pipelines and data flows, Using source data as Data lake data from Dataverse and Sink as dataset from the Azure SQL Database but this process is taking time. More than 9mins to sink 32 tables. Each table contains less than 2000 rows on average. and One table contains 250k rows is there a way to reduce the time of this process?

 

Problem 2: 

We are have an issue with triggering the pipelines. Is there any way to trigger on data change?

If any data is changed in the data verse then the pipe line should trigger and update the respective record or table in the Azure SQL DB.

 

Please help me in this.

 

 

 

 

 

1 Reply

  • Your slow performance is likely because you are using Data Flows, which spin up Spark and add overhead for small-to-medium datasets. For faster loads, switch to Copy Data Activity instead of Data Flows, enable bulk insert/PolyBase for Azure SQL, and run multiple table copies in parallel instead of sequentially. This should bring the runtime down significantly.

    For triggering, ADF cannot detect Dataverse changes directly. The common approach is to use Dataverse Change Tracking with Power Automate or Logic Apps to call your ADF pipeline whenever data changes, so only the updated rows get synced instead of reloading everything.

Resources