Compare Azure Data Factory V1 and V2

This article compares V2 with V1 of Azure Data Factory. For an introduction to V1, see Introduction to Azure Data Factory. For an introduction to V2, see Introduction to Data Factory (V2 - preview).

Feature comparison

The following table compares features of V1 and V2.

Feature Version 1 Version 2
Datasets A named view of data that references the data that you want to use in your activities as inputs and outputs. Datasets identify data within different data stores, such as tables, files, folders, and documents. For example, an Azure Blob dataset specifies the blob container and folder in Azure Blob storage from which the activity should read the data.

Availability defines the processing window slicing model for the dataset (for example, hourly, daily, and so on).
Datasets are the same in V2. However, you do not need to define availability schedules for datasets. You can define a trigger resource that can schedule pipelines from a clock scheduler paradigm. For more information, see Triggers and Datasets.
Linked services Linked services are much like connection strings, which define the connection information that's necessary for Data Factory to connect to external resources. Linked services are the same as in Data Factory V1, but with a new connectVia property to utilize the Data Factory V2 Integration Runtime compute environment. For more information, see Integration runtime in Azure Data Factory and Linked service properties for Azure Blob storage.
Pipelines A data factory can have one or more pipelines. A pipeline is a logical grouping of activities that together perform a task. You use startTime, endTime, and isPaused to schedule and run pipelines. Pipelines are groups of activities that are performed on data. However, the scheduling of activities in the pipeline has been separated into new trigger resources. You can think of pipelines in Data Factory V2 more as “workflow units” that you schedule separately via triggers.

Pipelines do not have “windows” of time execution in Data Factory V2. The Data Factory V1 concepts of startTime, endTime, and isPaused are no longer present in Data Factory V2. For more information, see Pipeline execution and triggers and Pipelines and activities.
Activities Activities define actions to perform on your data within a pipeline. Data movement (copy activity) and data transformation activities (such as Hive, Pig, and MapReduce) are supported. In Data Factory V2, activities still are defined actions within a pipeline. V2 introduces new control flow activities. You use these activities in a control flow (looping and branching). Data movement and data transformation activities that were supported in V1 are supported in V2. You can define transformation activities without using datasets in V2.
Hybrid data movement and activity dispatch Now called Integration Runtime, Data Management Gateway supported moving data between on-premises and cloud. Data Management Gateway is now called Self-Hosted Integration Runtime. It provides the same capability as it did in V1.

The Azure-SSIS Integration Runtime in V2 also supports deploying and running SQL Server Integration Services (SSIS) packages in the cloud. For more information, see Integration runtime in Azure Data Factory.
Parameters NA Parameters are key-value pairs of read-only configuration settings that are defined in pipelines. You can pass arguments for the parameters when you are manually running the pipeline. If you are using a scheduler trigger, the trigger can pass values for the parameters too. Activities within the pipeline consume the parameter values.
Expressions Data Factory V1 allows you to use functions and system variables in data selection queries and activity/dataset properties. In Data Factory V2, you can use expressions anywhere in a JSON string value. For more information, see Expressions and functions in V2.
Pipeline runs NA A single instance of a pipeline execution. For example, say you have a pipeline that executes at 8 AM, 9 AM, and 10 AM. There would be three separate runs of the pipeline (pipeline runs) in this case. Each pipeline run has a unique pipeline run ID. The pipeline run ID is a GUID that uniquely defines that particular pipeline run. Pipeline runs are typically instantiated by passing arguments to parameters that are defined in the pipelines.
Activity runs NA An instance of an activity execution within a pipeline.
Trigger runs NA An instance of a trigger execution. For more information, see Triggers.
Scheduling Scheduling is based on pipeline start/end times and dataset availability. Scheduler trigger or execution via external scheduler. For more information, see Pipeline execution and triggers.

The following sections provide more information about the capabilities of V2.

Control flow

To support diverse integration flows and patterns in the modern data warehouse, Data Factory V2 has enabled a new flexible data pipeline model that is no longer tied to time-series data. A few common flows that were previously not possible are now enabled. They are described in the following sections.

Chaining activities

In V1, you had to configure the output of an activity as an input of another activity to chain them. In V2, you can chain activities in a sequence within a pipeline. You can use the dependsOn property in an activity definition to chain it with an upstream activity. For more information and an example, see Pipelines and activities and Branching and chaining activities.

Branching activities

In V2, you can branch activities within a pipeline. The If-condition activity provides the same functionality that an if statement provides in programming languages. It evaluates a set of activities when the condition evaluates to true and another set of activities when the condition evaluates to false. For examples of branching activities, see the Branching and chaining activities tutorial.


You can define parameters at the pipeline level and pass arguments while you're invoking the pipeline on-demand or from a trigger. Activities can consume the arguments that are passed to the pipeline. For more information, see Pipelines and triggers.

Custom state passing

Activity outputs including state can be consumed by a subsequent activity in the pipeline. For example, in the JSON definition of an activity, you can access the output of the previous activity by using the following syntax: @activity('NameofPreviousActivity').output.value. By using this feature, you can build workflows where values can pass through activities.

Looping containers

The ForEach activity defines a repeating control flow in your pipeline. This activity iterates over a collection and runs specified activities in a loop. The loop implementation of this activity is similar to the Foreach looping structure in programming languages.

The Until activity provides the same functionality that a do-until looping structure provides in programming languages. It runs a set of activities in a loop until the condition that's associated with the activity evaluates to true. You can specify a timeout value for the until activity in Data Factory.

Trigger-based flows

Pipelines can be triggered by on-demand or wall-clock time. The pipelines and triggers article has detailed information about triggers.

Invoking a pipeline from another pipeline

The Execute Pipeline activity allows a Data Factory pipeline to invoke another pipeline.

Delta flows

A key use case in ETL patterns is “delta loads,” in which only data that has changed since the last iteration of a pipeline is loaded. New capabilities in V2, such as lookup activity, flexible scheduling, and control flow, enable this use case in a natural way. For a tutorial with step-by-step instructions, see Tutorial: Incremental copy.

Other control flow activities

Following are a few more control flow activities that are supported by Data Factory V2.

Control activity Description
ForEach activity Defines a repeating control flow in your pipeline. This activity is used to iterate over a collection and runs specified activities in a loop. The loop implementation of this activity is similar to Foreach looping structure in programming languages.
Web activity Calls a custom REST endpoint from a Data Factory pipeline. You can pass datasets and linked services to be consumed and accessed by the activity.
Lookup activity Reads or looks up a record or table name value from any external source. This output can further be referenced by succeeding activities.
Get metadata activity Retrieves the metadata of any data in Azure Data Factory.
Wait activity Pauses the pipeline for a specified period of time.

Deploy SSIS packages to Azure

You use Azure-SSIS if you want to move your SSIS workloads to the cloud, create a data factory by using V2, and provision an Azure-SSIS Integration Runtime.

The Azure-SSIS Integration Runtime is a fully managed cluster of Azure VMs (nodes) that are dedicated to running your SSIS packages in the cloud. After you provision Azure-SSIS Integration Runtime, you can use the same tools that you have been using to deploy SSIS packages to an on-premises SSIS environment.

For example, you can use SQL Server Data Tools or SQL Server Management Studio to deploy SSIS packages to this runtime on Azure. For step-by-step instructions, see the tutorial Deploy SQL Server integration services packages to Azure.

Flexible scheduling

In Data Factory V2, you do not need to define dataset availability schedules. You can define a trigger resource that can schedule pipelines from a clock scheduler paradigm. You can also pass parameters to pipelines from a trigger for a flexible scheduling and execution model.

Pipelines do not have “windows” of time execution in Data Factory V2. The Data Factory V1 concepts of startTime, endTime, and isPaused don't exist in Data Factory V2. For more information about how to build and then schedule a pipeline in Data Factory V2, see Pipeline execution and triggers.

Support for more data stores

V2 supports the copying of data to and from more data stores than V1. For a list of supported data stores, see the following articles:

Support for on-demand Spark cluster

V2 supports the creation of an on-demand Azure HDInsight Spark cluster. To create an on-demand Spark cluster, specify the cluster type as Spark in your on-demand, HDInsight linked service definition. Then you can configure the Spark activity in your pipeline to use this linked service.

At runtime, when the activity is executed, the Data Factory service automatically creates the Spark cluster for you. For more information, see the following articles:

Custom activities

In V1, you implement (custom) DotNet activity code by creating a .NET class library project with a class that implements the Execute method of the IDotNetActivity interface. Therefore, you need to write your custom code in .NET Framework 4.5.2 and run it on Windows-based Azure Batch Pool nodes.

In a V2 custom activity, you don't have to implement a .NET interface. You can directly run commands, scripts, and your own custom code compiled as an executable.

For more information, see Difference between custom activity in V1 and V2.


Data Factory V2 provides a richer set of SDKs that can be used to author, manage, and monitor pipelines.

  • .NET SDK: The .NET SDK is updated for V2.

  • PowerShell: The PowerShell cmdlets are updated for V2. The V2 cmdlets have DataFactoryV2 in the name, for example: Get-AzureRmDataFactoryV2.

  • Python SDK: This SDK is new to V2.

  • REST API: The REST API is updated for V2.

The SDKs that are updated for V2 are not backward-compatible with V1 clients.

Authoring experience

  V2 V1
Azure portal Yes Yes
Azure PowerShell Yes Yes
.NET SDK Yes Yes
Python SDK Yes No
Resource Manager template Yes Yes

Roles and permissions

The Data Factory version 1 Contributor role can be used to create and manage Data Factory v2 resources. For more info, see Data Factory Contributor.

Monitoring experience

In V2, you can also monitor data factories by using Azure Monitor. The new PowerShell cmdlets support monitoring of integration runtimes. Both V1 and V2 support visual monitoring via a monitoring application that can be launched from the Azure portal.

Next steps

Learn how to create a data factory by following step-by-step instructions in the following quickstarts: PowerShell, .NET, Python, REST API.