New data flow features: Import schema and test connection from debug cluster, custom sink ordering
Published Feb 04 2020 03:16 PM 3,167 Views

Several new features were added to mapping data flows this past week. Here are some of the highlights:


Import Schema from debug cluster


You can now use an active debug cluster to create a schema projection in your data flow source. 

Available in every source type, importing the schema will override the projection defined in the dataset. The dataset object will not be changed. All previously existing methods of creating and modifying schemas are still valid and compatible.




For more information on the projection tab, see the data flow source documentation 


Test connection on Spark Cluster


You can use an active debug cluster to verify data factory can connect to your linked service when using Spark in data flows. This is useful as a sanity check to ensure your dataset and linked service are valid configurations when used in data flows. 


Custom sink ordering


If you have multiple destinations in your data flow, you can now specify the write order of your data flow. Nondeterministic by default, enabling custom sink ordering allows for sequential writes of your data flow sinks.




For more information, see custom sink ordering 

Version history
Last update:
‎Feb 04 2020 03:16 PM
Updated by: