Blog Post

Azure Data Factory Blog
1 MIN READ

Introducing Map Data preview for Synapse

JoshuhaOwen's avatar
JoshuhaOwen
Icon for Microsoft rankMicrosoft
Jan 13, 2022

This week, we are excited to announce the public preview for Map Data, a new feature for Azure Synapse Analytics and Database Templates! The Map Data tool is a guided process to help users create ETL mappings and mapping data flows from their source data to Synapse lake database tables without writing code. This experience will help you get started with transformations into your Synapse Lake database quickly but still give you the power of Mapping Data Flows.

 

This process starts with the user choosing the destination tables in Synapse lake databases and then mapping their source data into these tables. We will be following up with a demo video shortly.

 

The Map Data tool is started from within the Synapse lake database experience. From here, you can select the Map Data tool to begin the process.

 

 

You will be able to use a guided approach to define source-to-target mappings between your chosen sources and Lake Database tables.

 

The Map Data tool will quickly allow you to map your input files and sources to Lake Database tables, generate a mapping data flow, and quickly have data flowing in an easy-to-use experience. Please do share your feedback and we look forward to seeing how we can make things even easier for our community!

Updated Jan 13, 2022
Version 1.0
  • MrAndyCutler's avatar
    MrAndyCutler
    Copper Contributor

    Hi,

     

    Is this Append only?  The pipeline that is generated is a simple copy from/to, each time the pipeline is run the data is duplicated in the destination.  Are there any settings for incremental?

  • Ideally, more than one source should be selectable when doing a mapping: For example, when loading a fact table (main source: ADLS), the surrogate keys to the dimensions should be obtained from Azure LakeDB (second source: Azure LakeDB). At the moment the lookup to the dimensions has to be done manually after the pipeline generation in the data flow itself.