Microsoft.DataFactory factories/pipelines 2018-06-01

Bicep resource definition

The factories/pipelines resource type can be deployed with operations that target:

For a list of changed properties in each API version, see change log.

Resource format

To create a Microsoft.DataFactory/factories/pipelines resource, add the following Bicep to your template.

resource symbolicname 'Microsoft.DataFactory/factories/pipelines@2018-06-01' = {
  name: 'string'
  parent: resourceSymbolicName
  properties: {
    activities: [
      {
        dependsOn: [
          {
            activity: 'string'
            dependencyConditions: [
              'string'
            ]
            {customized property}: any()
          }
        ]
        description: 'string'
        name: 'string'
        onInactiveMarkAs: 'string'
        state: 'string'
        userProperties: [
          {
            name: 'string'
            value: any()
          }
        ]
        type: 'string'
        // For remaining properties, see Activity objects
      }
    ]
    annotations: [
      any
    ]
    concurrency: int
    description: 'string'
    folder: {
      name: 'string'
    }
    parameters: {
      {customized property}: {
        defaultValue: any()
        type: 'string'
      }
    }
    policy: {
      elapsedTimeMetric: {
        duration: any()
      }
    }
    runDimensions: {
      {customized property}: any()
    }
    variables: {
      {customized property}: {
        defaultValue: any()
        type: 'string'
      }
    }
  }
}

Activity objects

Set the type property to specify the type of object.

For AppendVariable, use:

  type: 'AppendVariable'
  typeProperties: {
    value: any()
    variableName: 'string'
  }

For AzureDataExplorerCommand, use:

  type: 'AzureDataExplorerCommand'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    command: any()
    commandTimeout: any()
  }

For AzureFunctionActivity, use:

  type: 'AzureFunctionActivity'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    body: any()
    functionName: any()
    headers: {
      {customized property}: 'string'
    }
    method: 'string'
  }

For AzureMLBatchExecution, use:

  type: 'AzureMLBatchExecution'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    globalParameters: {
      {customized property}: any()
    }
    webServiceInputs: {
      {customized property}: {
        filePath: any()
        linkedServiceName: {
          parameters: {
            {customized property}: any()
          }
          referenceName: 'string'
          type: 'LinkedServiceReference'
        }
      }
    }
    webServiceOutputs: {
      {customized property}: {
        filePath: any()
        linkedServiceName: {
          parameters: {
            {customized property}: any()
          }
          referenceName: 'string'
          type: 'LinkedServiceReference'
        }
      }
    }
  }

For AzureMLExecutePipeline, use:

  type: 'AzureMLExecutePipeline'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    continueOnStepFailure: any()
    dataPathAssignments: any()
    experimentName: any()
    mlParentRunId: any()
    mlPipelineEndpointId: any()
    mlPipelineId: any()
    mlPipelineParameters: any()
    version: any()
  }

For AzureMLUpdateResource, use:

  type: 'AzureMLUpdateResource'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    trainedModelFilePath: any()
    trainedModelLinkedServiceName: {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'LinkedServiceReference'
    }
    trainedModelName: any()
  }

For Copy, use:

  type: 'Copy'
  inputs: [
    {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'DatasetReference'
    }
  ]
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  outputs: [
    {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'DatasetReference'
    }
  ]
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    dataIntegrationUnits: any()
    enableSkipIncompatibleRow: any()
    enableStaging: any()
    logSettings: {
      copyActivityLogSettings: {
        enableReliableLogging: any()
        logLevel: any()
      }
      enableCopyActivityLog: any()
      logLocationSettings: {
        linkedServiceName: {
          parameters: {
            {customized property}: any()
          }
          referenceName: 'string'
          type: 'LinkedServiceReference'
        }
        path: any()
      }
    }
    logStorageSettings: {
      enableReliableLogging: any()
      linkedServiceName: {
        parameters: {
          {customized property}: any()
        }
        referenceName: 'string'
        type: 'LinkedServiceReference'
      }
      logLevel: any()
      path: any()
      {customized property}: any()
    }
    parallelCopies: any()
    preserve: [
      any
    ]
    preserveRules: [
      any
    ]
    redirectIncompatibleRowSettings: {
      linkedServiceName: any()
      path: any()
      {customized property}: any()
    }
    sink: {
      disableMetricsCollection: any()
      maxConcurrentConnections: any()
      sinkRetryCount: any()
      sinkRetryWait: any()
      writeBatchSize: any()
      writeBatchTimeout: any()
      type: 'string'
      // For remaining properties, see CopySink objects
    }
    skipErrorFile: {
      dataInconsistency: any()
      fileMissing: any()
    }
    source: {
      disableMetricsCollection: any()
      maxConcurrentConnections: any()
      sourceRetryCount: any()
      sourceRetryWait: any()
      type: 'string'
      // For remaining properties, see CopySource objects
    }
    stagingSettings: {
      enableCompression: any()
      linkedServiceName: {
        parameters: {
          {customized property}: any()
        }
        referenceName: 'string'
        type: 'LinkedServiceReference'
      }
      path: any()
      {customized property}: any()
    }
    translator: any()
    validateDataConsistency: any()
  }

For Custom, use:

  type: 'Custom'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    autoUserSpecification: any()
    command: any()
    extendedProperties: {
      {customized property}: any()
    }
    folderPath: any()
    referenceObjects: {
      datasets: [
        {
          parameters: {
            {customized property}: any()
          }
          referenceName: 'string'
          type: 'DatasetReference'
        }
      ]
      linkedServices: [
        {
          parameters: {
            {customized property}: any()
          }
          referenceName: 'string'
          type: 'LinkedServiceReference'
        }
      ]
    }
    resourceLinkedService: {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'LinkedServiceReference'
    }
    retentionTimeInDays: any()
  }

For DatabricksNotebook, use:

  type: 'DatabricksNotebook'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    baseParameters: {
      {customized property}: any()
    }
    libraries: [
      {
        {customized property}: any()
      }
    ]
    notebookPath: any()
  }

For DatabricksSparkJar, use:

  type: 'DatabricksSparkJar'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    libraries: [
      {
        {customized property}: any()
      }
    ]
    mainClassName: any()
    parameters: [
      any
    ]
  }

For DatabricksSparkPython, use:

  type: 'DatabricksSparkPython'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    libraries: [
      {
        {customized property}: any()
      }
    ]
    parameters: [
      any
    ]
    pythonFile: any()
  }

For DataLakeAnalyticsU-SQL, use:

  type: 'DataLakeAnalyticsU-SQL'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    compilationMode: any()
    degreeOfParallelism: any()
    parameters: {
      {customized property}: any()
    }
    priority: any()
    runtimeVersion: any()
    scriptLinkedService: {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'LinkedServiceReference'
    }
    scriptPath: any()
  }

For Delete, use:

  type: 'Delete'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    dataset: {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'DatasetReference'
    }
    enableLogging: any()
    logStorageSettings: {
      enableReliableLogging: any()
      linkedServiceName: {
        parameters: {
          {customized property}: any()
        }
        referenceName: 'string'
        type: 'LinkedServiceReference'
      }
      logLevel: any()
      path: any()
      {customized property}: any()
    }
    maxConcurrentConnections: int
    recursive: any()
    storeSettings: {
      disableMetricsCollection: any()
      maxConcurrentConnections: any()
      type: 'string'
      // For remaining properties, see StoreReadSettings objects
    }
  }

For ExecuteDataFlow, use:

  type: 'ExecuteDataFlow'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    compute: {
      computeType: any()
      coreCount: any()
    }
    continueOnError: any()
    dataFlow: {
      datasetParameters: any()
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'DataFlowReference'
      {customized property}: any()
    }
    integrationRuntime: {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'IntegrationRuntimeReference'
    }
    runConcurrently: any()
    sourceStagingConcurrency: any()
    staging: {
      folderPath: any()
      linkedService: {
        parameters: {
          {customized property}: any()
        }
        referenceName: 'string'
        type: 'LinkedServiceReference'
      }
    }
    traceLevel: any()
  }

For ExecutePipeline, use:

  type: 'ExecutePipeline'
  policy: {
    secureInput: bool
    {customized property}: any()
  }
  typeProperties: {
    parameters: {
      {customized property}: any()
    }
    pipeline: {
      name: 'string'
      referenceName: 'string'
      type: 'PipelineReference'
    }
    waitOnCompletion: bool
  }

For ExecuteSSISPackage, use:

  type: 'ExecuteSSISPackage'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    connectVia: {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'IntegrationRuntimeReference'
    }
    environmentPath: any()
    executionCredential: {
      domain: any()
      password: {
        type: 'string'
        value: 'string'
      }
      userName: any()
    }
    loggingLevel: any()
    logLocation: {
      logPath: any()
      type: 'File'
      typeProperties: {
        accessCredential: {
          domain: any()
          password: {
            type: 'string'
            // For remaining properties, see SecretBase objects
          }
          userName: any()
        }
        logRefreshInterval: any()
      }
    }
    packageConnectionManagers: {
      {customized property}: {
        {customized property}: {
          value: any()
        }
      }
    }
    packageLocation: {
      packagePath: any()
      type: 'string'
      typeProperties: {
        accessCredential: {
          domain: any()
          password: {
            type: 'string'
            // For remaining properties, see SecretBase objects
          }
          userName: any()
        }
        childPackages: [
          {
            packageContent: any()
            packageLastModifiedDate: 'string'
            packageName: 'string'
            packagePath: any()
          }
        ]
        configurationAccessCredential: {
          domain: any()
          password: {
            type: 'string'
            // For remaining properties, see SecretBase objects
          }
          userName: any()
        }
        configurationPath: any()
        packageContent: any()
        packageLastModifiedDate: 'string'
        packageName: 'string'
        packagePassword: {
          type: 'string'
          // For remaining properties, see SecretBase objects
        }
      }
    }
    packageParameters: {
      {customized property}: {
        value: any()
      }
    }
    projectConnectionManagers: {
      {customized property}: {
        {customized property}: {
          value: any()
        }
      }
    }
    projectParameters: {
      {customized property}: {
        value: any()
      }
    }
    propertyOverrides: {
      {customized property}: {
        isSensitive: bool
        value: any()
      }
    }
    runtime: any()
  }

For ExecuteWranglingDataflow, use:

  type: 'ExecuteWranglingDataflow'
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    compute: {
      computeType: any()
      coreCount: any()
    }
    continueOnError: any()
    dataFlow: {
      datasetParameters: any()
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'DataFlowReference'
      {customized property}: any()
    }
    integrationRuntime: {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'IntegrationRuntimeReference'
    }
    queries: [
      {
        dataflowSinks: [
          {
            dataset: {
              parameters: {
                {customized property}: any()
              }
              referenceName: 'string'
              type: 'DatasetReference'
            }
            description: 'string'
            flowlet: {
              datasetParameters: any()
              parameters: {
                {customized property}: any()
              }
              referenceName: 'string'
              type: 'DataFlowReference'
              {customized property}: any()
            }
            linkedService: {
              parameters: {
                {customized property}: any()
              }
              referenceName: 'string'
              type: 'LinkedServiceReference'
            }
            name: 'string'
            rejectedDataLinkedService: {
              parameters: {
                {customized property}: any()
              }
              referenceName: 'string'
              type: 'LinkedServiceReference'
            }
            schemaLinkedService: {
              parameters: {
                {customized property}: any()
              }
              referenceName: 'string'
              type: 'LinkedServiceReference'
            }
            script: 'string'
          }
        ]
        queryName: 'string'
      }
    ]
    runConcurrently: any()
    sinks: {
      {customized property}: {
        dataset: {
          parameters: {
            {customized property}: any()
          }
          referenceName: 'string'
          type: 'DatasetReference'
        }
        description: 'string'
        flowlet: {
          datasetParameters: any()
          parameters: {
            {customized property}: any()
          }
          referenceName: 'string'
          type: 'DataFlowReference'
          {customized property}: any()
        }
        linkedService: {
          parameters: {
            {customized property}: any()
          }
          referenceName: 'string'
          type: 'LinkedServiceReference'
        }
        name: 'string'
        rejectedDataLinkedService: {
          parameters: {
            {customized property}: any()
          }
          referenceName: 'string'
          type: 'LinkedServiceReference'
        }
        schemaLinkedService: {
          parameters: {
            {customized property}: any()
          }
          referenceName: 'string'
          type: 'LinkedServiceReference'
        }
        script: 'string'
      }
    }
    sourceStagingConcurrency: any()
    staging: {
      folderPath: any()
      linkedService: {
        parameters: {
          {customized property}: any()
        }
        referenceName: 'string'
        type: 'LinkedServiceReference'
      }
    }
    traceLevel: any()
  }

For Fail, use:

  type: 'Fail'
  typeProperties: {
    errorCode: any()
    message: any()
  }

For Filter, use:

  type: 'Filter'
  typeProperties: {
    condition: {
      type: 'Expression'
      value: 'string'
    }
    items: {
      type: 'Expression'
      value: 'string'
    }
  }

For ForEach, use:

  type: 'ForEach'
  typeProperties: {
    activities: [
      {
        dependsOn: [
          {
            activity: 'string'
            dependencyConditions: [
              'string'
            ]
            {customized property}: any()
          }
        ]
        description: 'string'
        name: 'string'
        onInactiveMarkAs: 'string'
        state: 'string'
        userProperties: [
          {
            name: 'string'
            value: any()
          }
        ]
        type: 'string'
        // For remaining properties, see Activity objects
      }
    ]
    batchCount: int
    isSequential: bool
    items: {
      type: 'Expression'
      value: 'string'
    }
  }

For GetMetadata, use:

  type: 'GetMetadata'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    dataset: {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'DatasetReference'
    }
    fieldList: [
      any
    ]
    formatSettings: {
      type: 'string'
      // For remaining properties, see FormatReadSettings objects
    }
    storeSettings: {
      disableMetricsCollection: any()
      maxConcurrentConnections: any()
      type: 'string'
      // For remaining properties, see StoreReadSettings objects
    }
  }

For HDInsightHive, use:

  type: 'HDInsightHive'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    arguments: [
      any
    ]
    defines: {
      {customized property}: any()
    }
    getDebugInfo: 'string'
    queryTimeout: int
    scriptLinkedService: {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'LinkedServiceReference'
    }
    scriptPath: any()
    storageLinkedServices: [
      {
        parameters: {
          {customized property}: any()
        }
        referenceName: 'string'
        type: 'LinkedServiceReference'
      }
    ]
    variables: {
      {customized property}: any()
    }
  }

For HDInsightMapReduce, use:

  type: 'HDInsightMapReduce'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    arguments: [
      any
    ]
    className: any()
    defines: {
      {customized property}: any()
    }
    getDebugInfo: 'string'
    jarFilePath: any()
    jarLibs: [
      any
    ]
    jarLinkedService: {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'LinkedServiceReference'
    }
    storageLinkedServices: [
      {
        parameters: {
          {customized property}: any()
        }
        referenceName: 'string'
        type: 'LinkedServiceReference'
      }
    ]
  }

For HDInsightPig, use:

  type: 'HDInsightPig'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    arguments: any()
    defines: {
      {customized property}: any()
    }
    getDebugInfo: 'string'
    scriptLinkedService: {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'LinkedServiceReference'
    }
    scriptPath: any()
    storageLinkedServices: [
      {
        parameters: {
          {customized property}: any()
        }
        referenceName: 'string'
        type: 'LinkedServiceReference'
      }
    ]
  }

For HDInsightSpark, use:

  type: 'HDInsightSpark'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    arguments: [
      any
    ]
    className: 'string'
    entryFilePath: any()
    getDebugInfo: 'string'
    proxyUser: any()
    rootPath: any()
    sparkConfig: {
      {customized property}: any()
    }
    sparkJobLinkedService: {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'LinkedServiceReference'
    }
  }

For HDInsightStreaming, use:

  type: 'HDInsightStreaming'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    arguments: [
      any
    ]
    combiner: any()
    commandEnvironment: [
      any
    ]
    defines: {
      {customized property}: any()
    }
    fileLinkedService: {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'LinkedServiceReference'
    }
    filePaths: [
      any
    ]
    getDebugInfo: 'string'
    input: any()
    mapper: any()
    output: any()
    reducer: any()
    storageLinkedServices: [
      {
        parameters: {
          {customized property}: any()
        }
        referenceName: 'string'
        type: 'LinkedServiceReference'
      }
    ]
  }

For IfCondition, use:

  type: 'IfCondition'
  typeProperties: {
    expression: {
      type: 'Expression'
      value: 'string'
    }
    ifFalseActivities: [
      {
        dependsOn: [
          {
            activity: 'string'
            dependencyConditions: [
              'string'
            ]
            {customized property}: any()
          }
        ]
        description: 'string'
        name: 'string'
        onInactiveMarkAs: 'string'
        state: 'string'
        userProperties: [
          {
            name: 'string'
            value: any()
          }
        ]
        type: 'string'
        // For remaining properties, see Activity objects
      }
    ]
    ifTrueActivities: [
      {
        dependsOn: [
          {
            activity: 'string'
            dependencyConditions: [
              'string'
            ]
            {customized property}: any()
          }
        ]
        description: 'string'
        name: 'string'
        onInactiveMarkAs: 'string'
        state: 'string'
        userProperties: [
          {
            name: 'string'
            value: any()
          }
        ]
        type: 'string'
        // For remaining properties, see Activity objects
      }
    ]
  }

For Lookup, use:

  type: 'Lookup'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    dataset: {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'DatasetReference'
    }
    firstRowOnly: any()
    source: {
      disableMetricsCollection: any()
      maxConcurrentConnections: any()
      sourceRetryCount: any()
      sourceRetryWait: any()
      type: 'string'
      // For remaining properties, see CopySource objects
    }
  }

For Script, use:

  type: 'Script'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    logSettings: {
      logDestination: 'string'
      logLocationSettings: {
        linkedServiceName: {
          parameters: {
            {customized property}: any()
          }
          referenceName: 'string'
          type: 'LinkedServiceReference'
        }
        path: any()
      }
    }
    scriptBlockExecutionTimeout: any()
    scripts: [
      {
        parameters: [
          {
            direction: 'string'
            name: any()
            size: int
            type: 'string'
            value: any()
          }
        ]
        text: any()
        type: 'string'
      }
    ]
  }

For SetVariable, use:

  type: 'SetVariable'
  policy: {
    secureInput: bool
    secureOutput: bool
  }
  typeProperties: {
    setSystemVariable: bool
    value: any()
    variableName: 'string'
  }

For SparkJob, use:

  type: 'SparkJob'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    args: [
      any
    ]
    className: any()
    conf: any()
    configurationType: 'string'
    driverSize: any()
    executorSize: any()
    file: any()
    files: [
      any
    ]
    filesV2: [
      any
    ]
    numExecutors: any()
    pythonCodeReference: [
      any
    ]
    scanFolder: any()
    sparkConfig: {
      {customized property}: any()
    }
    sparkJob: {
      referenceName: any()
      type: 'SparkJobDefinitionReference'
    }
    targetBigDataPool: {
      referenceName: any()
      type: 'BigDataPoolReference'
    }
    targetSparkConfiguration: {
      referenceName: any()
      type: 'SparkConfigurationReference'
    }
  }

For SqlServerStoredProcedure, use:

  type: 'SqlServerStoredProcedure'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    storedProcedureName: any()
    storedProcedureParameters: any()
  }

For Switch, use:

  type: 'Switch'
  typeProperties: {
    cases: [
      {
        activities: [
          {
            dependsOn: [
              {
                activity: 'string'
                dependencyConditions: [
                  'string'
                ]
                {customized property}: any()
              }
            ]
            description: 'string'
            name: 'string'
            onInactiveMarkAs: 'string'
            state: 'string'
            userProperties: [
              {
                name: 'string'
                value: any()
              }
            ]
            type: 'string'
            // For remaining properties, see Activity objects
          }
        ]
        value: 'string'
      }
    ]
    defaultActivities: [
      {
        dependsOn: [
          {
            activity: 'string'
            dependencyConditions: [
              'string'
            ]
            {customized property}: any()
          }
        ]
        description: 'string'
        name: 'string'
        onInactiveMarkAs: 'string'
        state: 'string'
        userProperties: [
          {
            name: 'string'
            value: any()
          }
        ]
        type: 'string'
        // For remaining properties, see Activity objects
      }
    ]
    on: {
      type: 'Expression'
      value: 'string'
    }
  }

For SynapseNotebook, use:

  type: 'SynapseNotebook'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    conf: any()
    configurationType: 'string'
    driverSize: any()
    executorSize: any()
    notebook: {
      referenceName: any()
      type: 'NotebookReference'
    }
    numExecutors: any()
    parameters: {
      {customized property}: {
        type: 'string'
        value: any()
      }
    }
    sparkConfig: {
      {customized property}: any()
    }
    sparkPool: {
      referenceName: any()
      type: 'BigDataPoolReference'
    }
    targetSparkConfiguration: {
      referenceName: any()
      type: 'SparkConfigurationReference'
    }
  }

For Until, use:

  type: 'Until'
  typeProperties: {
    activities: [
      {
        dependsOn: [
          {
            activity: 'string'
            dependencyConditions: [
              'string'
            ]
            {customized property}: any()
          }
        ]
        description: 'string'
        name: 'string'
        onInactiveMarkAs: 'string'
        state: 'string'
        userProperties: [
          {
            name: 'string'
            value: any()
          }
        ]
        type: 'string'
        // For remaining properties, see Activity objects
      }
    ]
    expression: {
      type: 'Expression'
      value: 'string'
    }
    timeout: any()
  }

For Validation, use:

  type: 'Validation'
  typeProperties: {
    childItems: any()
    dataset: {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'DatasetReference'
    }
    minimumSize: any()
    sleep: any()
    timeout: any()
  }

For Wait, use:

  type: 'Wait'
  typeProperties: {
    waitTimeInSeconds: any()
  }

For WebActivity, use:

  type: 'WebActivity'
  linkedServiceName: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }
  policy: {
    retry: any()
    retryIntervalInSeconds: int
    secureInput: bool
    secureOutput: bool
    timeout: any()
    {customized property}: any()
  }
  typeProperties: {
    authentication: {
      credential: {
        referenceName: 'string'
        type: 'CredentialReference'
        {customized property}: any()
      }
      password: {
        type: 'string'
        // For remaining properties, see SecretBase objects
      }
      pfx: {
        type: 'string'
        // For remaining properties, see SecretBase objects
      }
      resource: any()
      type: 'string'
      username: any()
      userTenant: any()
    }
    body: any()
    connectVia: {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'IntegrationRuntimeReference'
    }
    datasets: [
      {
        parameters: {
          {customized property}: any()
        }
        referenceName: 'string'
        type: 'DatasetReference'
      }
    ]
    disableCertValidation: bool
    headers: {
      {customized property}: 'string'
    }
    httpRequestTimeout: any()
    linkedServices: [
      {
        parameters: {
          {customized property}: any()
        }
        referenceName: 'string'
        type: 'LinkedServiceReference'
      }
    ]
    method: 'string'
    turnOffAsync: bool
    url: any()
  }

For WebHook, use:

  type: 'WebHook'
  policy: {
    secureInput: bool
    secureOutput: bool
  }
  typeProperties: {
    authentication: {
      credential: {
        referenceName: 'string'
        type: 'CredentialReference'
        {customized property}: any()
      }
      password: {
        type: 'string'
        // For remaining properties, see SecretBase objects
      }
      pfx: {
        type: 'string'
        // For remaining properties, see SecretBase objects
      }
      resource: any()
      type: 'string'
      username: any()
      userTenant: any()
    }
    body: any()
    headers: {
      {customized property}: 'string'
    }
    method: 'POST'
    reportStatusOnCallBack: any()
    timeout: 'string'
    url: any()
  }

CopySink objects

Set the type property to specify the type of object.

For AvroSink, use:

  type: 'AvroSink'
  formatSettings: {
    fileNamePrefix: any()
    maxRowsPerFile: any()
    recordName: 'string'
    recordNamespace: 'string'
    type: 'string'
    {customized property}: any()
  }
  storeSettings: {
    copyBehavior: any()
    disableMetricsCollection: any()
    maxConcurrentConnections: any()
    metadata: [
      {
        name: any()
        value: any()
      }
    ]
    type: 'string'
    // For remaining properties, see StoreWriteSettings objects
  }

For AzureBlobFSSink, use:

  type: 'AzureBlobFSSink'
  copyBehavior: any()
  metadata: [
    {
      name: any()
      value: any()
    }
  ]

For AzureDatabricksDeltaLakeSink, use:

  type: 'AzureDatabricksDeltaLakeSink'
  importSettings: {
    dateFormat: any()
    timestampFormat: any()
    type: 'string'
    {customized property}: any()
  }
  preCopyScript: any()

For AzureDataExplorerSink, use:

  type: 'AzureDataExplorerSink'
  flushImmediately: any()
  ingestionMappingAsJson: any()
  ingestionMappingName: any()

For AzureDataLakeStoreSink, use:

  type: 'AzureDataLakeStoreSink'
  copyBehavior: any()
  enableAdlsSingleFileParallel: any()

For AzureMySqlSink, use:

  type: 'AzureMySqlSink'
  preCopyScript: any()

For AzurePostgreSqlSink, use:

  type: 'AzurePostgreSqlSink'
  preCopyScript: any()

For AzureQueueSink, use:

  type: 'AzureQueueSink'

For AzureSearchIndexSink, use:

  type: 'AzureSearchIndexSink'
  writeBehavior: 'string'

For AzureSqlSink, use:

  type: 'AzureSqlSink'
  preCopyScript: any()
  sqlWriterStoredProcedureName: any()
  sqlWriterTableType: any()
  sqlWriterUseTableLock: any()
  storedProcedureParameters: any()
  storedProcedureTableTypeParameterName: any()
  tableOption: any()
  upsertSettings: {
    interimSchemaName: any()
    keys: any()
    useTempDB: any()
  }
  writeBehavior: any()

For AzureTableSink, use:

  type: 'AzureTableSink'
  azureTableDefaultPartitionKeyValue: any()
  azureTableInsertType: any()
  azureTablePartitionKeyName: any()
  azureTableRowKeyName: any()

For BinarySink, use:

  type: 'BinarySink'
  storeSettings: {
    copyBehavior: any()
    disableMetricsCollection: any()
    maxConcurrentConnections: any()
    metadata: [
      {
        name: any()
        value: any()
      }
    ]
    type: 'string'
    // For remaining properties, see StoreWriteSettings objects
  }

For BlobSink, use:

  type: 'BlobSink'
  blobWriterAddHeader: any()
  blobWriterDateTimeFormat: any()
  blobWriterOverwriteFiles: any()
  copyBehavior: any()
  metadata: [
    {
      name: any()
      value: any()
    }
  ]

For CommonDataServiceForAppsSink, use:

  type: 'CommonDataServiceForAppsSink'
  alternateKeyName: any()
  ignoreNullValues: any()
  writeBehavior: 'Upsert'

For CosmosDbMongoDbApiSink, use:

  type: 'CosmosDbMongoDbApiSink'
  writeBehavior: any()

For CosmosDbSqlApiSink, use:

  type: 'CosmosDbSqlApiSink'
  writeBehavior: any()

For DelimitedTextSink, use:

  type: 'DelimitedTextSink'
  formatSettings: {
    fileExtension: any()
    fileNamePrefix: any()
    maxRowsPerFile: any()
    quoteAllText: any()
    type: 'string'
    {customized property}: any()
  }
  storeSettings: {
    copyBehavior: any()
    disableMetricsCollection: any()
    maxConcurrentConnections: any()
    metadata: [
      {
        name: any()
        value: any()
      }
    ]
    type: 'string'
    // For remaining properties, see StoreWriteSettings objects
  }

For DocumentDbCollectionSink, use:

  type: 'DocumentDbCollectionSink'
  nestingSeparator: any()
  writeBehavior: any()

For DynamicsCrmSink, use:

  type: 'DynamicsCrmSink'
  alternateKeyName: any()
  ignoreNullValues: any()
  writeBehavior: 'Upsert'

For DynamicsSink, use:

  type: 'DynamicsSink'
  alternateKeyName: any()
  ignoreNullValues: any()
  writeBehavior: 'Upsert'

For FileSystemSink, use:

  type: 'FileSystemSink'
  copyBehavior: any()

For InformixSink, use:

  type: 'InformixSink'
  preCopyScript: any()

For JsonSink, use:

  type: 'JsonSink'
  formatSettings: {
    filePattern: any()
    type: 'string'
    {customized property}: any()
  }
  storeSettings: {
    copyBehavior: any()
    disableMetricsCollection: any()
    maxConcurrentConnections: any()
    metadata: [
      {
        name: any()
        value: any()
      }
    ]
    type: 'string'
    // For remaining properties, see StoreWriteSettings objects
  }

For LakeHouseTableSink, use:

  type: 'LakeHouseTableSink'
  partitionNameList: any()
  partitionOption: any()
  tableActionOption: any()

For MicrosoftAccessSink, use:

  type: 'MicrosoftAccessSink'
  preCopyScript: any()

For MongoDbAtlasSink, use:

  type: 'MongoDbAtlasSink'
  writeBehavior: any()

For MongoDbV2Sink, use:

  type: 'MongoDbV2Sink'
  writeBehavior: any()

For OdbcSink, use:

  type: 'OdbcSink'
  preCopyScript: any()

For OracleSink, use:

  type: 'OracleSink'
  preCopyScript: any()

For OrcSink, use:

  type: 'OrcSink'
  formatSettings: {
    fileNamePrefix: any()
    maxRowsPerFile: any()
    type: 'string'
    {customized property}: any()
  }
  storeSettings: {
    copyBehavior: any()
    disableMetricsCollection: any()
    maxConcurrentConnections: any()
    metadata: [
      {
        name: any()
        value: any()
      }
    ]
    type: 'string'
    // For remaining properties, see StoreWriteSettings objects
  }

For ParquetSink, use:

  type: 'ParquetSink'
  formatSettings: {
    fileNamePrefix: any()
    maxRowsPerFile: any()
    type: 'string'
    {customized property}: any()
  }
  storeSettings: {
    copyBehavior: any()
    disableMetricsCollection: any()
    maxConcurrentConnections: any()
    metadata: [
      {
        name: any()
        value: any()
      }
    ]
    type: 'string'
    // For remaining properties, see StoreWriteSettings objects
  }

For RestSink, use:

  type: 'RestSink'
  additionalHeaders: any()
  httpCompressionType: any()
  httpRequestTimeout: any()
  requestInterval: any()
  requestMethod: any()

For SalesforceServiceCloudSink, use:

  type: 'SalesforceServiceCloudSink'
  externalIdFieldName: any()
  ignoreNullValues: any()
  writeBehavior: 'string'

For SalesforceServiceCloudV2Sink, use:

  type: 'SalesforceServiceCloudV2Sink'
  externalIdFieldName: any()
  ignoreNullValues: any()
  writeBehavior: 'string'

For SalesforceSink, use:

  type: 'SalesforceSink'
  externalIdFieldName: any()
  ignoreNullValues: any()
  writeBehavior: 'string'

For SalesforceV2Sink, use:

  type: 'SalesforceV2Sink'
  externalIdFieldName: any()
  ignoreNullValues: any()
  writeBehavior: 'string'

For SapCloudForCustomerSink, use:

  type: 'SapCloudForCustomerSink'
  httpRequestTimeout: any()
  writeBehavior: 'string'

For SnowflakeSink, use:

  type: 'SnowflakeSink'
  importSettings: {
    additionalCopyOptions: {
      {customized property}: any()
    }
    additionalFormatOptions: {
      {customized property}: any()
    }
    type: 'string'
    {customized property}: any()
  }
  preCopyScript: any()

For SnowflakeV2Sink, use:

  type: 'SnowflakeV2Sink'
  importSettings: {
    additionalCopyOptions: {
      {customized property}: any()
    }
    additionalFormatOptions: {
      {customized property}: any()
    }
    type: 'string'
    {customized property}: any()
  }
  preCopyScript: any()

For SqlDWSink, use:

  type: 'SqlDWSink'
  allowCopyCommand: any()
  allowPolyBase: any()
  copyCommandSettings: {
    additionalOptions: {
      {customized property}: 'string'
    }
    defaultValues: [
      {
        columnName: any()
        defaultValue: any()
      }
    ]
  }
  polyBaseSettings: {
    rejectSampleValue: any()
    rejectType: 'string'
    rejectValue: any()
    useTypeDefault: any()
    {customized property}: any()
  }
  preCopyScript: any()
  sqlWriterUseTableLock: any()
  tableOption: any()
  upsertSettings: {
    interimSchemaName: any()
    keys: any()
  }
  writeBehavior: any()

For SqlMISink, use:

  type: 'SqlMISink'
  preCopyScript: any()
  sqlWriterStoredProcedureName: any()
  sqlWriterTableType: any()
  sqlWriterUseTableLock: any()
  storedProcedureParameters: any()
  storedProcedureTableTypeParameterName: any()
  tableOption: any()
  upsertSettings: {
    interimSchemaName: any()
    keys: any()
    useTempDB: any()
  }
  writeBehavior: any()

For SqlServerSink, use:

  type: 'SqlServerSink'
  preCopyScript: any()
  sqlWriterStoredProcedureName: any()
  sqlWriterTableType: any()
  sqlWriterUseTableLock: any()
  storedProcedureParameters: any()
  storedProcedureTableTypeParameterName: any()
  tableOption: any()
  upsertSettings: {
    interimSchemaName: any()
    keys: any()
    useTempDB: any()
  }
  writeBehavior: any()

For SqlSink, use:

  type: 'SqlSink'
  preCopyScript: any()
  sqlWriterStoredProcedureName: any()
  sqlWriterTableType: any()
  sqlWriterUseTableLock: any()
  storedProcedureParameters: any()
  storedProcedureTableTypeParameterName: any()
  tableOption: any()
  upsertSettings: {
    interimSchemaName: any()
    keys: any()
    useTempDB: any()
  }
  writeBehavior: any()

For WarehouseSink, use:

  type: 'WarehouseSink'
  allowCopyCommand: any()
  copyCommandSettings: {
    additionalOptions: {
      {customized property}: 'string'
    }
    defaultValues: [
      {
        columnName: any()
        defaultValue: any()
      }
    ]
  }
  preCopyScript: any()
  tableOption: any()
  writeBehavior: any()

StoreWriteSettings objects

Set the type property to specify the type of object.

For AzureBlobFSWriteSettings, use:

  type: 'AzureBlobFSWriteSettings'
  blockSizeInMB: any()

For AzureBlobStorageWriteSettings, use:

  type: 'AzureBlobStorageWriteSettings'
  blockSizeInMB: any()

For AzureDataLakeStoreWriteSettings, use:

  type: 'AzureDataLakeStoreWriteSettings'
  expiryDateTime: any()

For AzureFileStorageWriteSettings, use:

  type: 'AzureFileStorageWriteSettings'

For FileServerWriteSettings, use:

  type: 'FileServerWriteSettings'

For LakeHouseWriteSettings, use:

  type: 'LakeHouseWriteSettings'

For SftpWriteSettings, use:

  type: 'SftpWriteSettings'
  operationTimeout: any()
  useTempFileRename: any()

CopySource objects

Set the type property to specify the type of object.

For AmazonMWSSource, use:

  type: 'AmazonMWSSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For AmazonRdsForOracleSource, use:

  type: 'AmazonRdsForOracleSource'
  additionalColumns: any()
  oracleReaderQuery: any()
  partitionOption: any()
  partitionSettings: {
    partitionColumnName: any()
    partitionLowerBound: any()
    partitionNames: any()
    partitionUpperBound: any()
  }
  queryTimeout: any()

For AmazonRdsForSqlServerSource, use:

  type: 'AmazonRdsForSqlServerSource'
  additionalColumns: any()
  isolationLevel: any()
  partitionOption: any()
  partitionSettings: {
    partitionColumnName: any()
    partitionLowerBound: any()
    partitionUpperBound: any()
  }
  produceAdditionalTypes: any()
  queryTimeout: any()
  sqlReaderQuery: any()
  sqlReaderStoredProcedureName: any()
  storedProcedureParameters: any()

For AmazonRedshiftSource, use:

  type: 'AmazonRedshiftSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()
  redshiftUnloadSettings: {
    bucketName: any()
    s3LinkedServiceName: {
      parameters: {
        {customized property}: any()
      }
      referenceName: 'string'
      type: 'LinkedServiceReference'
    }
  }

For AvroSource, use:

  type: 'AvroSource'
  additionalColumns: any()
  storeSettings: {
    disableMetricsCollection: any()
    maxConcurrentConnections: any()
    type: 'string'
    // For remaining properties, see StoreReadSettings objects
  }

For AzureBlobFSSource, use:

  type: 'AzureBlobFSSource'
  recursive: any()
  skipHeaderLineCount: any()
  treatEmptyAsNull: any()

For AzureDatabricksDeltaLakeSource, use:

  type: 'AzureDatabricksDeltaLakeSource'
  exportSettings: {
    dateFormat: any()
    timestampFormat: any()
    type: 'string'
    {customized property}: any()
  }
  query: any()

For AzureDataExplorerSource, use:

  type: 'AzureDataExplorerSource'
  additionalColumns: any()
  noTruncation: any()
  query: any()
  queryTimeout: any()

For AzureDataLakeStoreSource, use:

  type: 'AzureDataLakeStoreSource'
  recursive: any()

For AzureMariaDBSource, use:

  type: 'AzureMariaDBSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For AzureMySqlSource, use:

  type: 'AzureMySqlSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For AzurePostgreSqlSource, use:

  type: 'AzurePostgreSqlSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For AzureSqlSource, use:

  type: 'AzureSqlSource'
  additionalColumns: any()
  isolationLevel: any()
  partitionOption: any()
  partitionSettings: {
    partitionColumnName: any()
    partitionLowerBound: any()
    partitionUpperBound: any()
  }
  produceAdditionalTypes: any()
  queryTimeout: any()
  sqlReaderQuery: any()
  sqlReaderStoredProcedureName: any()
  storedProcedureParameters: any()

For AzureTableSource, use:

  type: 'AzureTableSource'
  additionalColumns: any()
  azureTableSourceIgnoreTableNotFound: any()
  azureTableSourceQuery: any()
  queryTimeout: any()

For BinarySource, use:

  type: 'BinarySource'
  formatSettings: {
    compressionProperties: {
      type: 'string'
      // For remaining properties, see CompressionReadSettings objects
    }
    type: 'string'
    {customized property}: any()
  }
  storeSettings: {
    disableMetricsCollection: any()
    maxConcurrentConnections: any()
    type: 'string'
    // For remaining properties, see StoreReadSettings objects
  }

For BlobSource, use:

  type: 'BlobSource'
  recursive: any()
  skipHeaderLineCount: any()
  treatEmptyAsNull: any()

For CassandraSource, use:

  type: 'CassandraSource'
  additionalColumns: any()
  consistencyLevel: 'string'
  query: any()
  queryTimeout: any()

For CommonDataServiceForAppsSource, use:

  type: 'CommonDataServiceForAppsSource'
  additionalColumns: any()
  query: any()

For ConcurSource, use:

  type: 'ConcurSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For CosmosDbMongoDbApiSource, use:

  type: 'CosmosDbMongoDbApiSource'
  additionalColumns: any()
  batchSize: any()
  cursorMethods: {
    limit: any()
    project: any()
    skip: any()
    sort: any()
    {customized property}: any()
  }
  filter: any()
  queryTimeout: any()

For CosmosDbSqlApiSource, use:

  type: 'CosmosDbSqlApiSource'
  additionalColumns: any()
  detectDatetime: any()
  pageSize: any()
  preferredRegions: any()
  query: any()

For CouchbaseSource, use:

  type: 'CouchbaseSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For Db2Source, use:

  type: 'Db2Source'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For DelimitedTextSource, use:

  type: 'DelimitedTextSource'
  additionalColumns: any()
  formatSettings: {
    compressionProperties: {
      type: 'string'
      // For remaining properties, see CompressionReadSettings objects
    }
    skipLineCount: any()
    type: 'string'
    {customized property}: any()
  }
  storeSettings: {
    disableMetricsCollection: any()
    maxConcurrentConnections: any()
    type: 'string'
    // For remaining properties, see StoreReadSettings objects
  }

For DocumentDbCollectionSource, use:

  type: 'DocumentDbCollectionSource'
  additionalColumns: any()
  nestingSeparator: any()
  query: any()
  queryTimeout: any()

For DrillSource, use:

  type: 'DrillSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For DynamicsAXSource, use:

  type: 'DynamicsAXSource'
  additionalColumns: any()
  httpRequestTimeout: any()
  query: any()
  queryTimeout: any()

For DynamicsCrmSource, use:

  type: 'DynamicsCrmSource'
  additionalColumns: any()
  query: any()

For DynamicsSource, use:

  type: 'DynamicsSource'
  additionalColumns: any()
  query: any()

For EloquaSource, use:

  type: 'EloquaSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For ExcelSource, use:

  type: 'ExcelSource'
  additionalColumns: any()
  storeSettings: {
    disableMetricsCollection: any()
    maxConcurrentConnections: any()
    type: 'string'
    // For remaining properties, see StoreReadSettings objects
  }

For FileSystemSource, use:

  type: 'FileSystemSource'
  additionalColumns: any()
  recursive: any()

For GoogleAdWordsSource, use:

  type: 'GoogleAdWordsSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For GoogleBigQuerySource, use:

  type: 'GoogleBigQuerySource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For GoogleBigQueryV2Source, use:

  type: 'GoogleBigQueryV2Source'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For GreenplumSource, use:

  type: 'GreenplumSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For HBaseSource, use:

  type: 'HBaseSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For HdfsSource, use:

  type: 'HdfsSource'
  distcpSettings: {
    distcpOptions: any()
    resourceManagerEndpoint: any()
    tempScriptPath: any()
  }
  recursive: any()

For HiveSource, use:

  type: 'HiveSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For HttpSource, use:

  type: 'HttpSource'
  httpRequestTimeout: any()

For HubspotSource, use:

  type: 'HubspotSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For ImpalaSource, use:

  type: 'ImpalaSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For InformixSource, use:

  type: 'InformixSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For JiraSource, use:

  type: 'JiraSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For JsonSource, use:

  type: 'JsonSource'
  additionalColumns: any()
  formatSettings: {
    compressionProperties: {
      type: 'string'
      // For remaining properties, see CompressionReadSettings objects
    }
    type: 'string'
    {customized property}: any()
  }
  storeSettings: {
    disableMetricsCollection: any()
    maxConcurrentConnections: any()
    type: 'string'
    // For remaining properties, see StoreReadSettings objects
  }

For LakeHouseTableSource, use:

  type: 'LakeHouseTableSource'
  additionalColumns: any()
  timestampAsOf: any()
  versionAsOf: any()

For MagentoSource, use:

  type: 'MagentoSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For MariaDBSource, use:

  type: 'MariaDBSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For MarketoSource, use:

  type: 'MarketoSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For MicrosoftAccessSource, use:

  type: 'MicrosoftAccessSource'
  additionalColumns: any()
  query: any()

For MongoDbAtlasSource, use:

  type: 'MongoDbAtlasSource'
  additionalColumns: any()
  batchSize: any()
  cursorMethods: {
    limit: any()
    project: any()
    skip: any()
    sort: any()
    {customized property}: any()
  }
  filter: any()
  queryTimeout: any()

For MongoDbSource, use:

  type: 'MongoDbSource'
  additionalColumns: any()
  query: any()

For MongoDbV2Source, use:

  type: 'MongoDbV2Source'
  additionalColumns: any()
  batchSize: any()
  cursorMethods: {
    limit: any()
    project: any()
    skip: any()
    sort: any()
    {customized property}: any()
  }
  filter: any()
  queryTimeout: any()

For MySqlSource, use:

  type: 'MySqlSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For NetezzaSource, use:

  type: 'NetezzaSource'
  additionalColumns: any()
  partitionOption: any()
  partitionSettings: {
    partitionColumnName: any()
    partitionLowerBound: any()
    partitionUpperBound: any()
  }
  query: any()
  queryTimeout: any()

For ODataSource, use:

  type: 'ODataSource'
  additionalColumns: any()
  httpRequestTimeout: any()
  query: any()

For OdbcSource, use:

  type: 'OdbcSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For Office365Source, use:

  type: 'Office365Source'
  allowedGroups: any()
  dateFilterColumn: any()
  endTime: any()
  outputColumns: any()
  startTime: any()
  userScopeFilterUri: any()

For OracleServiceCloudSource, use:

  type: 'OracleServiceCloudSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For OracleSource, use:

  type: 'OracleSource'
  additionalColumns: any()
  oracleReaderQuery: any()
  partitionOption: any()
  partitionSettings: {
    partitionColumnName: any()
    partitionLowerBound: any()
    partitionNames: any()
    partitionUpperBound: any()
  }
  queryTimeout: any()

For OrcSource, use:

  type: 'OrcSource'
  additionalColumns: any()
  storeSettings: {
    disableMetricsCollection: any()
    maxConcurrentConnections: any()
    type: 'string'
    // For remaining properties, see StoreReadSettings objects
  }

For ParquetSource, use:

  type: 'ParquetSource'
  additionalColumns: any()
  formatSettings: {
    compressionProperties: {
      type: 'string'
      // For remaining properties, see CompressionReadSettings objects
    }
    type: 'string'
    {customized property}: any()
  }
  storeSettings: {
    disableMetricsCollection: any()
    maxConcurrentConnections: any()
    type: 'string'
    // For remaining properties, see StoreReadSettings objects
  }

For PaypalSource, use:

  type: 'PaypalSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For PhoenixSource, use:

  type: 'PhoenixSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For PostgreSqlSource, use:

  type: 'PostgreSqlSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For PostgreSqlV2Source, use:

  type: 'PostgreSqlV2Source'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For PrestoSource, use:

  type: 'PrestoSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For QuickBooksSource, use:

  type: 'QuickBooksSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For RelationalSource, use:

  type: 'RelationalSource'
  additionalColumns: any()
  query: any()

For ResponsysSource, use:

  type: 'ResponsysSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For RestSource, use:

  type: 'RestSource'
  additionalColumns: any()
  additionalHeaders: any()
  httpRequestTimeout: any()
  paginationRules: any()
  requestBody: any()
  requestInterval: any()
  requestMethod: any()

For SalesforceMarketingCloudSource, use:

  type: 'SalesforceMarketingCloudSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For SalesforceServiceCloudSource, use:

  type: 'SalesforceServiceCloudSource'
  additionalColumns: any()
  query: any()
  readBehavior: any()

For SalesforceServiceCloudV2Source, use:

  type: 'SalesforceServiceCloudV2Source'
  additionalColumns: any()
  includeDeletedObjects: any()
  SOQLQuery: any()

For SalesforceSource, use:

  type: 'SalesforceSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()
  readBehavior: any()

For SalesforceV2Source, use:

  type: 'SalesforceV2Source'
  additionalColumns: any()
  includeDeletedObjects: any()
  queryTimeout: any()
  SOQLQuery: any()

For SapBwSource, use:

  type: 'SapBwSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For SapCloudForCustomerSource, use:

  type: 'SapCloudForCustomerSource'
  additionalColumns: any()
  httpRequestTimeout: any()
  query: any()
  queryTimeout: any()

For SapEccSource, use:

  type: 'SapEccSource'
  additionalColumns: any()
  httpRequestTimeout: any()
  query: any()
  queryTimeout: any()

For SapHanaSource, use:

  type: 'SapHanaSource'
  additionalColumns: any()
  packetSize: any()
  partitionOption: any()
  partitionSettings: {
    partitionColumnName: any()
  }
  query: any()
  queryTimeout: any()

For SapOdpSource, use:

  type: 'SapOdpSource'
  additionalColumns: any()
  extractionMode: any()
  projection: any()
  queryTimeout: any()
  selection: any()
  subscriberProcess: any()

For SapOpenHubSource, use:

  type: 'SapOpenHubSource'
  additionalColumns: any()
  baseRequestId: any()
  customRfcReadTableFunctionModule: any()
  excludeLastRequest: any()
  queryTimeout: any()
  sapDataColumnDelimiter: any()

For SapTableSource, use:

  type: 'SapTableSource'
  additionalColumns: any()
  batchSize: any()
  customRfcReadTableFunctionModule: any()
  partitionOption: any()
  partitionSettings: {
    maxPartitionsNumber: any()
    partitionColumnName: any()
    partitionLowerBound: any()
    partitionUpperBound: any()
  }
  queryTimeout: any()
  rfcTableFields: any()
  rfcTableOptions: any()
  rowCount: any()
  rowSkips: any()
  sapDataColumnDelimiter: any()

For ServiceNowSource, use:

  type: 'ServiceNowSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For ServiceNowV2Source, use:

  type: 'ServiceNowV2Source'
  additionalColumns: any()
  expression: {
    operands: [
      {
      operator: 'string'
      type: 'string'
      value: 'string'
  }
  queryTimeout: any()

For SharePointOnlineListSource, use:

  type: 'SharePointOnlineListSource'
  httpRequestTimeout: any()
  query: any()

For ShopifySource, use:

  type: 'ShopifySource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For SnowflakeSource, use:

  type: 'SnowflakeSource'
  exportSettings: {
    additionalCopyOptions: {
      {customized property}: any()
    }
    additionalFormatOptions: {
      {customized property}: any()
    }
    type: 'string'
    {customized property}: any()
  }
  query: any()

For SnowflakeV2Source, use:

  type: 'SnowflakeV2Source'
  exportSettings: {
    additionalCopyOptions: {
      {customized property}: any()
    }
    additionalFormatOptions: {
      {customized property}: any()
    }
    type: 'string'
    {customized property}: any()
  }
  query: any()

For SparkSource, use:

  type: 'SparkSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For SqlDWSource, use:

  type: 'SqlDWSource'
  additionalColumns: any()
  isolationLevel: any()
  partitionOption: any()
  partitionSettings: {
    partitionColumnName: any()
    partitionLowerBound: any()
    partitionUpperBound: any()
  }
  queryTimeout: any()
  sqlReaderQuery: any()
  sqlReaderStoredProcedureName: any()
  storedProcedureParameters: any()

For SqlMISource, use:

  type: 'SqlMISource'
  additionalColumns: any()
  isolationLevel: any()
  partitionOption: any()
  partitionSettings: {
    partitionColumnName: any()
    partitionLowerBound: any()
    partitionUpperBound: any()
  }
  produceAdditionalTypes: any()
  queryTimeout: any()
  sqlReaderQuery: any()
  sqlReaderStoredProcedureName: any()
  storedProcedureParameters: any()

For SqlServerSource, use:

  type: 'SqlServerSource'
  additionalColumns: any()
  isolationLevel: any()
  partitionOption: any()
  partitionSettings: {
    partitionColumnName: any()
    partitionLowerBound: any()
    partitionUpperBound: any()
  }
  produceAdditionalTypes: any()
  queryTimeout: any()
  sqlReaderQuery: any()
  sqlReaderStoredProcedureName: any()
  storedProcedureParameters: any()

For SqlSource, use:

  type: 'SqlSource'
  additionalColumns: any()
  isolationLevel: any()
  partitionOption: any()
  partitionSettings: {
    partitionColumnName: any()
    partitionLowerBound: any()
    partitionUpperBound: any()
  }
  queryTimeout: any()
  sqlReaderQuery: any()
  sqlReaderStoredProcedureName: any()
  storedProcedureParameters: any()

For SquareSource, use:

  type: 'SquareSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For SybaseSource, use:

  type: 'SybaseSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For TeradataSource, use:

  type: 'TeradataSource'
  additionalColumns: any()
  partitionOption: any()
  partitionSettings: {
    partitionColumnName: any()
    partitionLowerBound: any()
    partitionUpperBound: any()
  }
  query: any()
  queryTimeout: any()

For VerticaSource, use:

  type: 'VerticaSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For WarehouseSource, use:

  type: 'WarehouseSource'
  additionalColumns: any()
  isolationLevel: any()
  partitionOption: any()
  partitionSettings: {
    partitionColumnName: any()
    partitionLowerBound: any()
    partitionUpperBound: any()
  }
  queryTimeout: any()
  sqlReaderQuery: any()
  sqlReaderStoredProcedureName: any()
  storedProcedureParameters: any()

For WebSource, use:

  type: 'WebSource'
  additionalColumns: any()

For XeroSource, use:

  type: 'XeroSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

For XmlSource, use:

  type: 'XmlSource'
  additionalColumns: any()
  formatSettings: {
    compressionProperties: {
      type: 'string'
      // For remaining properties, see CompressionReadSettings objects
    }
    detectDataType: any()
    namespacePrefixes: any()
    namespaces: any()
    type: 'string'
    validationMode: any()
    {customized property}: any()
  }
  storeSettings: {
    disableMetricsCollection: any()
    maxConcurrentConnections: any()
    type: 'string'
    // For remaining properties, see StoreReadSettings objects
  }

For ZohoSource, use:

  type: 'ZohoSource'
  additionalColumns: any()
  query: any()
  queryTimeout: any()

StoreReadSettings objects

Set the type property to specify the type of object.

For AmazonS3CompatibleReadSettings, use:

  type: 'AmazonS3CompatibleReadSettings'
  deleteFilesAfterCompletion: any()
  enablePartitionDiscovery: any()
  fileListPath: any()
  modifiedDatetimeEnd: any()
  modifiedDatetimeStart: any()
  partitionRootPath: any()
  prefix: any()
  recursive: any()
  wildcardFileName: any()
  wildcardFolderPath: any()

For AmazonS3ReadSettings, use:

  type: 'AmazonS3ReadSettings'
  deleteFilesAfterCompletion: any()
  enablePartitionDiscovery: any()
  fileListPath: any()
  modifiedDatetimeEnd: any()
  modifiedDatetimeStart: any()
  partitionRootPath: any()
  prefix: any()
  recursive: any()
  wildcardFileName: any()
  wildcardFolderPath: any()

For AzureBlobFSReadSettings, use:

  type: 'AzureBlobFSReadSettings'
  deleteFilesAfterCompletion: any()
  enablePartitionDiscovery: any()
  fileListPath: any()
  modifiedDatetimeEnd: any()
  modifiedDatetimeStart: any()
  partitionRootPath: any()
  recursive: any()
  wildcardFileName: any()
  wildcardFolderPath: any()

For AzureBlobStorageReadSettings, use:

  type: 'AzureBlobStorageReadSettings'
  deleteFilesAfterCompletion: any()
  enablePartitionDiscovery: any()
  fileListPath: any()
  modifiedDatetimeEnd: any()
  modifiedDatetimeStart: any()
  partitionRootPath: any()
  prefix: any()
  recursive: any()
  wildcardFileName: any()
  wildcardFolderPath: any()

For AzureDataLakeStoreReadSettings, use:

  type: 'AzureDataLakeStoreReadSettings'
  deleteFilesAfterCompletion: any()
  enablePartitionDiscovery: any()
  fileListPath: any()
  listAfter: any()
  listBefore: any()
  modifiedDatetimeEnd: any()
  modifiedDatetimeStart: any()
  partitionRootPath: any()
  recursive: any()
  wildcardFileName: any()
  wildcardFolderPath: any()

For AzureFileStorageReadSettings, use:

  type: 'AzureFileStorageReadSettings'
  deleteFilesAfterCompletion: any()
  enablePartitionDiscovery: any()
  fileListPath: any()
  modifiedDatetimeEnd: any()
  modifiedDatetimeStart: any()
  partitionRootPath: any()
  prefix: any()
  recursive: any()
  wildcardFileName: any()
  wildcardFolderPath: any()

For FileServerReadSettings, use:

  type: 'FileServerReadSettings'
  deleteFilesAfterCompletion: any()
  enablePartitionDiscovery: any()
  fileFilter: any()
  fileListPath: any()
  modifiedDatetimeEnd: any()
  modifiedDatetimeStart: any()
  partitionRootPath: any()
  recursive: any()
  wildcardFileName: any()
  wildcardFolderPath: any()

For FtpReadSettings, use:

  type: 'FtpReadSettings'
  deleteFilesAfterCompletion: any()
  disableChunking: any()
  enablePartitionDiscovery: any()
  fileListPath: any()
  partitionRootPath: any()
  recursive: any()
  useBinaryTransfer: any()
  wildcardFileName: any()
  wildcardFolderPath: any()

For GoogleCloudStorageReadSettings, use:

  type: 'GoogleCloudStorageReadSettings'
  deleteFilesAfterCompletion: any()
  enablePartitionDiscovery: any()
  fileListPath: any()
  modifiedDatetimeEnd: any()
  modifiedDatetimeStart: any()
  partitionRootPath: any()
  prefix: any()
  recursive: any()
  wildcardFileName: any()
  wildcardFolderPath: any()

For HdfsReadSettings, use:

  type: 'HdfsReadSettings'
  deleteFilesAfterCompletion: any()
  distcpSettings: {
    distcpOptions: any()
    resourceManagerEndpoint: any()
    tempScriptPath: any()
  }
  enablePartitionDiscovery: any()
  fileListPath: any()
  modifiedDatetimeEnd: any()
  modifiedDatetimeStart: any()
  partitionRootPath: any()
  recursive: any()
  wildcardFileName: any()
  wildcardFolderPath: any()

For HttpReadSettings, use:

  type: 'HttpReadSettings'
  additionalColumns: any()
  additionalHeaders: any()
  requestBody: any()
  requestMethod: any()
  requestTimeout: any()

For LakeHouseReadSettings, use:

  type: 'LakeHouseReadSettings'
  deleteFilesAfterCompletion: any()
  enablePartitionDiscovery: any()
  fileListPath: any()
  modifiedDatetimeEnd: any()
  modifiedDatetimeStart: any()
  partitionRootPath: any()
  recursive: any()
  wildcardFileName: any()
  wildcardFolderPath: any()

For OracleCloudStorageReadSettings, use:

  type: 'OracleCloudStorageReadSettings'
  deleteFilesAfterCompletion: any()
  enablePartitionDiscovery: any()
  fileListPath: any()
  modifiedDatetimeEnd: any()
  modifiedDatetimeStart: any()
  partitionRootPath: any()
  prefix: any()
  recursive: any()
  wildcardFileName: any()
  wildcardFolderPath: any()

For SftpReadSettings, use:

  type: 'SftpReadSettings'
  deleteFilesAfterCompletion: any()
  disableChunking: any()
  enablePartitionDiscovery: any()
  fileListPath: any()
  modifiedDatetimeEnd: any()
  modifiedDatetimeStart: any()
  partitionRootPath: any()
  recursive: any()
  wildcardFileName: any()
  wildcardFolderPath: any()

CompressionReadSettings objects

Set the type property to specify the type of object.

For TarGZipReadSettings, use:

  type: 'TarGZipReadSettings'
  preserveCompressionFileNameAsFolder: any()

For TarReadSettings, use:

  type: 'TarReadSettings'
  preserveCompressionFileNameAsFolder: any()

For ZipDeflateReadSettings, use:

  type: 'ZipDeflateReadSettings'
  preserveZipFileNameAsFolder: any()

SecretBase objects

Set the type property to specify the type of object.

For AzureKeyVaultSecret, use:

  type: 'AzureKeyVaultSecret'
  secretName: any()
  secretVersion: any()
  store: {
    parameters: {
      {customized property}: any()
    }
    referenceName: 'string'
    type: 'LinkedServiceReference'
  }

For SecureString, use:

  type: 'SecureString'
  value: 'string'

FormatReadSettings objects

Set the type property to specify the type of object.

For BinaryReadSettings, use:

  type: 'BinaryReadSettings'
  compressionProperties: {
    type: 'string'
    // For remaining properties, see CompressionReadSettings objects
  }

For DelimitedTextReadSettings, use:

  type: 'DelimitedTextReadSettings'
  compressionProperties: {
    type: 'string'
    // For remaining properties, see CompressionReadSettings objects
  }
  skipLineCount: any()

For JsonReadSettings, use:

  type: 'JsonReadSettings'
  compressionProperties: {
    type: 'string'
    // For remaining properties, see CompressionReadSettings objects
  }

For ParquetReadSettings, use:

  type: 'ParquetReadSettings'
  compressionProperties: {
    type: 'string'
    // For remaining properties, see CompressionReadSettings objects
  }

For XmlReadSettings, use:

  type: 'XmlReadSettings'
  compressionProperties: {
    type: 'string'
    // For remaining properties, see CompressionReadSettings objects
  }
  detectDataType: any()
  namespacePrefixes: any()
  namespaces: any()
  validationMode: any()

Property values

factories/pipelines

Name Description Value
name The resource name

See how to set names and types for child resources in Bicep.
string (required)

Character limit: 1-260

Valid characters:
Can't use:
<>*#.%&:\\+?/ or control characters

Start with alphanumeric.
parent In Bicep, you can specify the parent resource for a child resource. You only need to add this property when the child resource is declared outside of the parent resource.

For more information, see Child resource outside parent resource.
Symbolic name for resource of type: factories
properties Properties of the pipeline. Pipeline (required)

Pipeline

Name Description Value
activities List of activities in pipeline. Activity[]
annotations List of tags that can be used for describing the Pipeline. any[]
concurrency The max number of concurrent runs for the pipeline. int

Constraints:
Min value = 1
description The description of the pipeline. string
folder The folder that this Pipeline is in. If not specified, Pipeline will appear at the root level. PipelineFolder
parameters List of parameters for pipeline. ParameterDefinitionSpecification
policy Pipeline Policy. PipelinePolicy
runDimensions Dimensions emitted by Pipeline. PipelineRunDimensions
variables List of variables for pipeline. VariableDefinitionSpecification

Activity

Name Description Value
dependsOn Activity depends on condition. ActivityDependency[]
description Activity description. string
name Activity name. string (required)
onInactiveMarkAs Status result of the activity when the state is set to Inactive. This is an optional property and if not provided when the activity is inactive, the status will be Succeeded by default. 'Failed'
'Skipped'
'Succeeded'
state Activity state. This is an optional property and if not provided, the state will be Active by default. 'Active'
'Inactive'
userProperties Activity user properties. UserProperty[]
type Set the object type AppendVariable
AzureDataExplorerCommand
AzureFunctionActivity
AzureMLBatchExecution
AzureMLExecutePipeline
AzureMLUpdateResource
Copy
Custom
DatabricksNotebook
DatabricksSparkJar
DatabricksSparkPython
DataLakeAnalyticsU-SQL
Delete
ExecuteDataFlow
ExecutePipeline
ExecuteSSISPackage
ExecuteWranglingDataflow
Fail
Filter
ForEach
GetMetadata
HDInsightHive
HDInsightMapReduce
HDInsightPig
HDInsightSpark
HDInsightStreaming
IfCondition
Lookup
Script
SetVariable
SparkJob
SqlServerStoredProcedure
Switch
SynapseNotebook
Until
Validation
Wait
WebActivity
WebHook (required)

ActivityDependency

Name Description Value
activity Activity name. string (required)
dependencyConditions Match-Condition for the dependency. String array containing any of:
'Completed'
'Failed'
'Skipped'
'Succeeded' (required)
{customized property} For Bicep, you can use the any() function.

UserProperty

Name Description Value
name User property name. string (required)
value User property value. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)

AppendVariableActivity

Name Description Value
type Type of activity. 'AppendVariable' (required)
typeProperties Append Variable activity properties. AppendVariableActivityTypeProperties (required)

AppendVariableActivityTypeProperties

Name Description Value
value Value to be appended. Type: could be a static value matching type of the variable item or Expression with resultType matching type of the variable item For Bicep, you can use the any() function.
variableName Name of the variable whose value needs to be appended to. string

AzureDataExplorerCommandActivity

Name Description Value
type Type of activity. 'AzureDataExplorerCommand' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Azure Data Explorer command activity properties. AzureDataExplorerCommandActivityTypeProperties (required)

LinkedServiceReference

Name Description Value
parameters Arguments for LinkedService. ParameterValueSpecification
referenceName Reference LinkedService name. string (required)
type Linked service reference type. 'LinkedServiceReference' (required)

ParameterValueSpecification

Name Description Value
{customized property} For Bicep, you can use the any() function.

ActivityPolicy

Name Description Value
retry Maximum ordinary retry attempts. Default is 0. Type: integer (or Expression with resultType integer), minimum: 0. For Bicep, you can use the any() function.
retryIntervalInSeconds Interval between each retry attempt (in seconds). The default is 30 sec. int

Constraints:
Min value = 30
Max value = 86400
secureInput When set to true, Input from activity is considered as secure and will not be logged to monitoring. bool
secureOutput When set to true, Output from activity is considered as secure and will not be logged to monitoring. bool
timeout Specifies the timeout for the activity to run. The default timeout is 7 days. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
{customized property} For Bicep, you can use the any() function.

AzureDataExplorerCommandActivityTypeProperties

Name Description Value
command A control command, according to the Azure Data Explorer command syntax. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
commandTimeout Control command timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9]))..) For Bicep, you can use the any() function.

AzureFunctionActivity

Name Description Value
type Type of activity. 'AzureFunctionActivity' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Azure Function activity properties. AzureFunctionActivityTypeProperties (required)

AzureFunctionActivityTypeProperties

Name Description Value
body Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
functionName Name of the Function that the Azure Function Activity will call. Type: string (or Expression with resultType string) For Bicep, you can use the any() function.(required)
headers Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). AzureFunctionActivityTypePropertiesHeaders
method Rest API method for target endpoint. 'DELETE'
'GET'
'HEAD'
'OPTIONS'
'POST'
'PUT'
'TRACE' (required)

AzureFunctionActivityTypePropertiesHeaders

Name Description Value
{customized property} string

AzureMLBatchExecutionActivity

Name Description Value
type Type of activity. 'AzureMLBatchExecution' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Azure ML Batch Execution activity properties. AzureMLBatchExecutionActivityTypeProperties (required)

AzureMLBatchExecutionActivityTypeProperties

Name Description Value
globalParameters Key,Value pairs to be passed to the Azure ML Batch Execution Service endpoint. Keys must match the names of web service parameters defined in the published Azure ML web service. Values will be passed in the GlobalParameters property of the Azure ML batch execution request. AzureMLBatchExecutionActivityTypePropertiesGlobalPar...
webServiceInputs Key,Value pairs, mapping the names of Azure ML endpoint's Web Service Inputs to AzureMLWebServiceFile objects specifying the input Blob locations.. This information will be passed in the WebServiceInputs property of the Azure ML batch execution request. AzureMLBatchExecutionActivityTypePropertiesWebServic...
webServiceOutputs Key,Value pairs, mapping the names of Azure ML endpoint's Web Service Outputs to AzureMLWebServiceFile objects specifying the output Blob locations. This information will be passed in the WebServiceOutputs property of the Azure ML batch execution request. AzureMLBatchExecutionActivityTypePropertiesWebServic...

AzureMLBatchExecutionActivityTypePropertiesGlobalPar...

Name Description Value
{customized property} For Bicep, you can use the any() function.

AzureMLBatchExecutionActivityTypePropertiesWebServic...

Name Description Value
{customized property} AzureMLWebServiceFile

AzureMLWebServiceFile

Name Description Value
filePath The relative file path, including container name, in the Azure Blob Storage specified by the LinkedService. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
linkedServiceName Reference to an Azure Storage LinkedService, where Azure ML WebService Input/Output file located. LinkedServiceReference (required)

AzureMLBatchExecutionActivityTypePropertiesWebServic...

Name Description Value
{customized property} AzureMLWebServiceFile

AzureMLExecutePipelineActivity

Name Description Value
type Type of activity. 'AzureMLExecutePipeline' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Azure ML Execute Pipeline activity properties. AzureMLExecutePipelineActivityTypeProperties (required)

AzureMLExecutePipelineActivityTypeProperties

Name Description Value
continueOnStepFailure Whether to continue execution of other steps in the PipelineRun if a step fails. This information will be passed in the continueOnStepFailure property of the published pipeline execution request. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
dataPathAssignments Dictionary used for changing data path assignments without retraining. Values will be passed in the dataPathAssignments property of the published pipeline execution request. Type: object (or Expression with resultType object). For Bicep, you can use the any() function.
experimentName Run history experiment name of the pipeline run. This information will be passed in the ExperimentName property of the published pipeline execution request. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
mlParentRunId The parent Azure ML Service pipeline run id. This information will be passed in the ParentRunId property of the published pipeline execution request. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
mlPipelineEndpointId ID of the published Azure ML pipeline endpoint. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
mlPipelineId ID of the published Azure ML pipeline. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
mlPipelineParameters Key,Value pairs to be passed to the published Azure ML pipeline endpoint. Keys must match the names of pipeline parameters defined in the published pipeline. Values will be passed in the ParameterAssignments property of the published pipeline execution request. Type: object with key value pairs (or Expression with resultType object). For Bicep, you can use the any() function.
version Version of the published Azure ML pipeline endpoint. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

AzureMLUpdateResourceActivity

Name Description Value
type Type of activity. 'AzureMLUpdateResource' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Azure ML Update Resource management activity properties. AzureMLUpdateResourceActivityTypeProperties (required)

AzureMLUpdateResourceActivityTypeProperties

Name Description Value
trainedModelFilePath The relative file path in trainedModelLinkedService to represent the .ilearner file that will be uploaded by the update operation. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
trainedModelLinkedServiceName Name of Azure Storage linked service holding the .ilearner file that will be uploaded by the update operation. LinkedServiceReference (required)
trainedModelName Name of the Trained Model module in the Web Service experiment to be updated. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)

CopyActivity

Name Description Value
type Type of activity. 'Copy' (required)
inputs List of inputs for the activity. DatasetReference[]
linkedServiceName Linked service reference. LinkedServiceReference
outputs List of outputs for the activity. DatasetReference[]
policy Activity policy. ActivityPolicy
typeProperties Copy activity properties. CopyActivityTypeProperties (required)

DatasetReference

Name Description Value
parameters Arguments for dataset. ParameterValueSpecification
referenceName Reference dataset name. string (required)
type Dataset reference type. 'DatasetReference' (required)

CopyActivityTypeProperties

Name Description Value
dataIntegrationUnits Maximum number of data integration units that can be used to perform this data movement. Type: integer (or Expression with resultType integer), minimum: 0. For Bicep, you can use the any() function.
enableSkipIncompatibleRow Whether to skip incompatible row. Default value is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
enableStaging Specifies whether to copy data via an interim staging. Default value is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
logSettings Log settings customer needs provide when enabling log. LogSettings
logStorageSettings (Deprecated. Please use LogSettings) Log storage settings customer need to provide when enabling session log. LogStorageSettings
parallelCopies Maximum number of concurrent sessions opened on the source or sink to avoid overloading the data store. Type: integer (or Expression with resultType integer), minimum: 0. For Bicep, you can use the any() function.
preserve Preserve rules. any[]
preserveRules Preserve Rules. any[]
redirectIncompatibleRowSettings Redirect incompatible row settings when EnableSkipIncompatibleRow is true. RedirectIncompatibleRowSettings
sink Copy activity sink. CopySink (required)
skipErrorFile Specify the fault tolerance for data consistency. SkipErrorFile
source Copy activity source. CopySource (required)
stagingSettings Specifies interim staging settings when EnableStaging is true. StagingSettings
translator Copy activity translator. If not specified, tabular translator is used. For Bicep, you can use the any() function.
validateDataConsistency Whether to enable Data Consistency validation. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.

LogSettings

Name Description Value
copyActivityLogSettings Specifies settings for copy activity log. CopyActivityLogSettings
enableCopyActivityLog Specifies whether to enable copy activity log. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
logLocationSettings Log location settings customer needs to provide when enabling log. LogLocationSettings (required)

CopyActivityLogSettings

Name Description Value
enableReliableLogging Specifies whether to enable reliable logging. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
logLevel Gets or sets the log level, support: Info, Warning. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

LogLocationSettings

Name Description Value
linkedServiceName Log storage linked service reference. LinkedServiceReference (required)
path The path to storage for storing detailed logs of activity execution. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

LogStorageSettings

Name Description Value
enableReliableLogging Specifies whether to enable reliable logging. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
linkedServiceName Log storage linked service reference. LinkedServiceReference (required)
logLevel Gets or sets the log level, support: Info, Warning. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
path The path to storage for storing detailed logs of activity execution. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
{customized property} For Bicep, you can use the any() function.

RedirectIncompatibleRowSettings

Name Description Value
linkedServiceName Name of the Azure Storage, Storage SAS, or Azure Data Lake Store linked service used for redirecting incompatible row. Must be specified if redirectIncompatibleRowSettings is specified. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
path The path for storing the redirect incompatible row data. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
{customized property} For Bicep, you can use the any() function.

CopySink

Name Description Value
disableMetricsCollection If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
maxConcurrentConnections The maximum concurrent connection count for the sink data store. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
sinkRetryCount Sink retry count. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
sinkRetryWait Sink retry wait. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
writeBatchSize Write batch size. Type: integer (or Expression with resultType integer), minimum: 0. For Bicep, you can use the any() function.
writeBatchTimeout Write batch timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
type Set the object type AvroSink
AzureBlobFSSink
AzureDatabricksDeltaLakeSink
AzureDataExplorerSink
AzureDataLakeStoreSink
AzureMySqlSink
AzurePostgreSqlSink
AzureQueueSink
AzureSearchIndexSink
AzureSqlSink
AzureTableSink
BinarySink
BlobSink
CommonDataServiceForAppsSink
CosmosDbMongoDbApiSink
CosmosDbSqlApiSink
DelimitedTextSink
DocumentDbCollectionSink
DynamicsCrmSink
DynamicsSink
FileSystemSink
InformixSink
JsonSink
LakeHouseTableSink
MicrosoftAccessSink
MongoDbAtlasSink
MongoDbV2Sink
OdbcSink
OracleSink
OrcSink
ParquetSink
RestSink
SalesforceServiceCloudSink
SalesforceServiceCloudV2Sink
SalesforceSink
SalesforceV2Sink
SapCloudForCustomerSink
SnowflakeSink
SnowflakeV2Sink
SqlDWSink
SqlMISink
SqlServerSink
SqlSink
WarehouseSink (required)

AvroSink

Name Description Value
type Copy sink type. 'AvroSink' (required)
formatSettings Avro format settings. AvroWriteSettings
storeSettings Avro store settings. StoreWriteSettings

AvroWriteSettings

Name Description Value
fileNamePrefix Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
maxRowsPerFile Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
recordName Top level record name in write result, which is required in AVRO spec. string
recordNamespace Record namespace in the write result. string
type The write setting type. string (required)
{customized property} For Bicep, you can use the any() function.

StoreWriteSettings

Name Description Value
copyBehavior The type of copy behavior for copy sink. For Bicep, you can use the any() function.
disableMetricsCollection If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
maxConcurrentConnections The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
metadata Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). MetadataItem[]
type Set the object type AzureBlobFSWriteSettings
AzureBlobStorageWriteSettings
AzureDataLakeStoreWriteSettings
AzureFileStorageWriteSettings
FileServerWriteSettings
LakeHouseWriteSettings
SftpWriteSettings (required)

MetadataItem

Name Description Value
name Metadata item key name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
value Metadata item value. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

AzureBlobFSWriteSettings

Name Description Value
type The write setting type. 'AzureBlobFSWriteSettings' (required)
blockSizeInMB Indicates the block size(MB) when writing data to blob. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.

AzureBlobStorageWriteSettings

Name Description Value
type The write setting type. 'AzureBlobStorageWriteSettings' (required)
blockSizeInMB Indicates the block size(MB) when writing data to blob. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.

AzureDataLakeStoreWriteSettings

Name Description Value
type The write setting type. 'AzureDataLakeStoreWriteSettings' (required)
expiryDateTime Specifies the expiry time of the written files. The time is applied to the UTC time zone in the format of "2018-12-01T05:00:00Z". Default value is NULL. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

AzureFileStorageWriteSettings

Name Description Value
type The write setting type. 'AzureFileStorageWriteSettings' (required)

FileServerWriteSettings

Name Description Value
type The write setting type. 'FileServerWriteSettings' (required)

LakeHouseWriteSettings

Name Description Value
type The write setting type. 'LakeHouseWriteSettings' (required)

SftpWriteSettings

Name Description Value
type The write setting type. 'SftpWriteSettings' (required)
operationTimeout Specifies the timeout for writing each chunk to SFTP server. Default value: 01:00:00 (one hour). Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
useTempFileRename Upload to temporary file(s) and rename. Disable this option if your SFTP server doesn't support rename operation. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.

AzureBlobFSSink

Name Description Value
type Copy sink type. 'AzureBlobFSSink' (required)
copyBehavior The type of copy behavior for copy sink. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
metadata Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). MetadataItem[]

AzureDatabricksDeltaLakeSink

Name Description Value
type Copy sink type. 'AzureDatabricksDeltaLakeSink' (required)
importSettings Azure Databricks Delta Lake import settings. AzureDatabricksDeltaLakeImportCommand
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

AzureDatabricksDeltaLakeImportCommand

Name Description Value
dateFormat Specify the date format for csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
timestampFormat Specify the timestamp format for csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
type The import setting type. string (required)
{customized property} For Bicep, you can use the any() function.

AzureDataExplorerSink

Name Description Value
type Copy sink type. 'AzureDataExplorerSink' (required)
flushImmediately If set to true, any aggregation will be skipped. Default is false. Type: boolean. For Bicep, you can use the any() function.
ingestionMappingAsJson An explicit column mapping description provided in a json format. Type: string. For Bicep, you can use the any() function.
ingestionMappingName A name of a pre-created csv mapping that was defined on the target Kusto table. Type: string. For Bicep, you can use the any() function.

AzureDataLakeStoreSink

Name Description Value
type Copy sink type. 'AzureDataLakeStoreSink' (required)
copyBehavior The type of copy behavior for copy sink. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
enableAdlsSingleFileParallel Single File Parallel. For Bicep, you can use the any() function.

AzureMySqlSink

Name Description Value
type Copy sink type. 'AzureMySqlSink' (required)
preCopyScript A query to execute before starting the copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

AzurePostgreSqlSink

Name Description Value
type Copy sink type. 'AzurePostgreSqlSink' (required)
preCopyScript A query to execute before starting the copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

AzureQueueSink

Name Description Value
type Copy sink type. 'AzureQueueSink' (required)

AzureSearchIndexSink

Name Description Value
type Copy sink type. 'AzureSearchIndexSink' (required)
writeBehavior Specify the write behavior when upserting documents into Azure Search Index. 'Merge'
'Upload'

AzureSqlSink

Name Description Value
type Copy sink type. 'AzureSqlSink' (required)
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlWriterStoredProcedureName SQL writer stored procedure name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlWriterTableType SQL writer table type. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlWriterUseTableLock Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
storedProcedureParameters SQL stored procedure parameters. For Bicep, you can use the any() function.
storedProcedureTableTypeParameterName The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
tableOption The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
upsertSettings SQL upsert settings. SqlUpsertSettings
writeBehavior Write behavior when copying data into Azure SQL. Type: SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum) For Bicep, you can use the any() function.

SqlUpsertSettings

Name Description Value
interimSchemaName Schema name for interim table. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
keys Key column names for unique row identification. Type: array of strings (or Expression with resultType array of strings). For Bicep, you can use the any() function.
useTempDB Specifies whether to use temp db for upsert interim table. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.

AzureTableSink

Name Description Value
type Copy sink type. 'AzureTableSink' (required)
azureTableDefaultPartitionKeyValue Azure Table default partition key value. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
azureTableInsertType Azure Table insert type. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
azureTablePartitionKeyName Azure Table partition key name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
azureTableRowKeyName Azure Table row key name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

BinarySink

Name Description Value
type Copy sink type. 'BinarySink' (required)
storeSettings Binary store settings. StoreWriteSettings

BlobSink

Name Description Value
type Copy sink type. 'BlobSink' (required)
blobWriterAddHeader Blob writer add header. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
blobWriterDateTimeFormat Blob writer date time format. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
blobWriterOverwriteFiles Blob writer overwrite files. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
copyBehavior The type of copy behavior for copy sink. For Bicep, you can use the any() function.
metadata Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). MetadataItem[]

CommonDataServiceForAppsSink

Name Description Value
type Copy sink type. 'CommonDataServiceForAppsSink' (required)
alternateKeyName The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
ignoreNullValues The flag indicating whether to ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
writeBehavior The write behavior for the operation. 'Upsert' (required)

CosmosDbMongoDbApiSink

Name Description Value
type Copy sink type. 'CosmosDbMongoDbApiSink' (required)
writeBehavior Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

CosmosDbSqlApiSink

Name Description Value
type Copy sink type. 'CosmosDbSqlApiSink' (required)
writeBehavior Describes how to write data to Azure Cosmos DB. Type: string (or Expression with resultType string). Allowed values: insert and upsert. For Bicep, you can use the any() function.

DelimitedTextSink

Name Description Value
type Copy sink type. 'DelimitedTextSink' (required)
formatSettings DelimitedText format settings. DelimitedTextWriteSettings
storeSettings DelimitedText store settings. StoreWriteSettings

DelimitedTextWriteSettings

Name Description Value
fileExtension The file extension used to create the files. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
fileNamePrefix Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
maxRowsPerFile Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
quoteAllText Indicates whether string values should always be enclosed with quotes. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
type The write setting type. string (required)
{customized property} For Bicep, you can use the any() function.

DocumentDbCollectionSink

Name Description Value
type Copy sink type. 'DocumentDbCollectionSink' (required)
nestingSeparator Nested properties separator. Default is . (dot). Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
writeBehavior Describes how to write data to Azure Cosmos DB. Type: string (or Expression with resultType string). Allowed values: insert and upsert. For Bicep, you can use the any() function.

DynamicsCrmSink

Name Description Value
type Copy sink type. 'DynamicsCrmSink' (required)
alternateKeyName The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
ignoreNullValues The flag indicating whether to ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
writeBehavior The write behavior for the operation. 'Upsert' (required)

DynamicsSink

Name Description Value
type Copy sink type. 'DynamicsSink' (required)
alternateKeyName The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
ignoreNullValues The flag indicating whether ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
writeBehavior The write behavior for the operation. 'Upsert' (required)

FileSystemSink

Name Description Value
type Copy sink type. 'FileSystemSink' (required)
copyBehavior The type of copy behavior for copy sink. For Bicep, you can use the any() function.

InformixSink

Name Description Value
type Copy sink type. 'InformixSink' (required)
preCopyScript A query to execute before starting the copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

JsonSink

Name Description Value
type Copy sink type. 'JsonSink' (required)
formatSettings Json format settings. JsonWriteSettings
storeSettings Json store settings. StoreWriteSettings

JsonWriteSettings

Name Description Value
filePattern File pattern of JSON. This setting controls the way a collection of JSON objects will be treated. The default value is 'setOfObjects'. It is case-sensitive. For Bicep, you can use the any() function.
type The write setting type. string (required)
{customized property} For Bicep, you can use the any() function.

LakeHouseTableSink

Name Description Value
type Copy sink type. 'LakeHouseTableSink' (required)
partitionNameList Specify the partition column names from sink columns. Type: array of objects (or Expression with resultType array of objects). For Bicep, you can use the any() function.
partitionOption Create partitions in folder structure based on one or multiple columns. Each distinct column value (pair) will be a new partition. Possible values include: "None", "PartitionByKey". For Bicep, you can use the any() function.
tableActionOption The type of table action for LakeHouse Table sink. Possible values include: "None", "Append", "Overwrite". For Bicep, you can use the any() function.

MicrosoftAccessSink

Name Description Value
type Copy sink type. 'MicrosoftAccessSink' (required)
preCopyScript A query to execute before starting the copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

MongoDbAtlasSink

Name Description Value
type Copy sink type. 'MongoDbAtlasSink' (required)
writeBehavior Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

MongoDbV2Sink

Name Description Value
type Copy sink type. 'MongoDbV2Sink' (required)
writeBehavior Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

OdbcSink

Name Description Value
type Copy sink type. 'OdbcSink' (required)
preCopyScript A query to execute before starting the copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

OracleSink

Name Description Value
type Copy sink type. 'OracleSink' (required)
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

OrcSink

Name Description Value
type Copy sink type. 'OrcSink' (required)
formatSettings ORC format settings. OrcWriteSettings
storeSettings ORC store settings. StoreWriteSettings

OrcWriteSettings

Name Description Value
fileNamePrefix Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
maxRowsPerFile Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
type The write setting type. string (required)
{customized property} For Bicep, you can use the any() function.

ParquetSink

Name Description Value
type Copy sink type. 'ParquetSink' (required)
formatSettings Parquet format settings. ParquetWriteSettings
storeSettings Parquet store settings. StoreWriteSettings

ParquetWriteSettings

Name Description Value
fileNamePrefix Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
maxRowsPerFile Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
type The write setting type. string (required)
{customized property} For Bicep, you can use the any() function.

RestSink

Name Description Value
type Copy sink type. 'RestSink' (required)
additionalHeaders The additional HTTP headers in the request to the RESTful API. Type: key value pairs (value should be string type). For Bicep, you can use the any() function.
httpCompressionType Http Compression Type to Send data in compressed format with Optimal Compression Level, Default is None. And The Only Supported option is Gzip. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
requestInterval The time to await before sending next request, in milliseconds For Bicep, you can use the any() function.
requestMethod The HTTP method used to call the RESTful API. The default is POST. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SalesforceServiceCloudSink

Name Description Value
type Copy sink type. 'SalesforceServiceCloudSink' (required)
externalIdFieldName The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
ignoreNullValues The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
writeBehavior The write behavior for the operation. Default is Insert. 'Insert'
'Upsert'

SalesforceServiceCloudV2Sink

Name Description Value
type Copy sink type. 'SalesforceServiceCloudV2Sink' (required)
externalIdFieldName The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
ignoreNullValues The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
writeBehavior The write behavior for the operation. Default is Insert. 'Insert'
'Upsert'

SalesforceSink

Name Description Value
type Copy sink type. 'SalesforceSink' (required)
externalIdFieldName The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
ignoreNullValues The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
writeBehavior The write behavior for the operation. Default is Insert. 'Insert'
'Upsert'

SalesforceV2Sink

Name Description Value
type Copy sink type. 'SalesforceV2Sink' (required)
externalIdFieldName The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
ignoreNullValues The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
writeBehavior The write behavior for the operation. Default is Insert. 'Insert'
'Upsert'

SapCloudForCustomerSink

Name Description Value
type Copy sink type. 'SapCloudForCustomerSink' (required)
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
writeBehavior The write behavior for the operation. Default is 'Insert'. 'Insert'
'Update'

SnowflakeSink

Name Description Value
type Copy sink type. 'SnowflakeSink' (required)
importSettings Snowflake import settings. SnowflakeImportCopyCommand
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SnowflakeImportCopyCommand

Name Description Value
additionalCopyOptions Additional copy options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": "'HH24:MI:SS.FF'" } SnowflakeImportCopyCommandAdditionalCopyOptions
additionalFormatOptions Additional format options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalFormatOptions": { "FORCE": "TRUE", "LOAD_UNCERTAIN_FILES": "'FALSE'" } SnowflakeImportCopyCommandAdditionalFormatOptions
type The import setting type. string (required)
{customized property} For Bicep, you can use the any() function.

SnowflakeImportCopyCommandAdditionalCopyOptions

Name Description Value
{customized property} For Bicep, you can use the any() function.

SnowflakeImportCopyCommandAdditionalFormatOptions

Name Description Value
{customized property} For Bicep, you can use the any() function.

SnowflakeV2Sink

Name Description Value
type Copy sink type. 'SnowflakeV2Sink' (required)
importSettings Snowflake import settings. SnowflakeImportCopyCommand
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SqlDWSink

Name Description Value
type Copy sink type. 'SqlDWSink' (required)
allowCopyCommand Indicates to use Copy Command to copy data into SQL Data Warehouse. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
allowPolyBase Indicates to use PolyBase to copy data into SQL Data Warehouse when applicable. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
copyCommandSettings Specifies Copy Command related settings when allowCopyCommand is true. DWCopyCommandSettings
polyBaseSettings Specifies PolyBase-related settings when allowPolyBase is true. PolybaseSettings
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlWriterUseTableLock Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
tableOption The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
upsertSettings SQL DW upsert settings. SqlDWUpsertSettings
writeBehavior Write behavior when copying data into azure SQL DW. Type: SqlDWWriteBehaviorEnum (or Expression with resultType SqlDWWriteBehaviorEnum) For Bicep, you can use the any() function.

DWCopyCommandSettings

Name Description Value
additionalOptions Additional options directly passed to SQL DW in Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalOptions": { "MAXERRORS": "1000", "DATEFORMAT": "'ymd'" } DWCopyCommandSettingsAdditionalOptions
defaultValues Specifies the default values for each target column in SQL DW. The default values in the property overwrite the DEFAULT constraint set in the DB, and identity column cannot have a default value. Type: array of objects (or Expression with resultType array of objects). DWCopyCommandDefaultValue[]

DWCopyCommandSettingsAdditionalOptions

Name Description Value
{customized property} string

DWCopyCommandDefaultValue

Name Description Value
columnName Column name. Type: object (or Expression with resultType string). For Bicep, you can use the any() function.
defaultValue The default value of the column. Type: object (or Expression with resultType string). For Bicep, you can use the any() function.

PolybaseSettings

Name Description Value
rejectSampleValue Determines the number of rows to attempt to retrieve before the PolyBase recalculates the percentage of rejected rows. Type: integer (or Expression with resultType integer), minimum: 0. For Bicep, you can use the any() function.
rejectType Reject type. 'percentage'
'value'
rejectValue Specifies the value or the percentage of rows that can be rejected before the query fails. Type: number (or Expression with resultType number), minimum: 0. For Bicep, you can use the any() function.
useTypeDefault Specifies how to handle missing values in delimited text files when PolyBase retrieves data from the text file. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
{customized property} For Bicep, you can use the any() function.

SqlDWUpsertSettings

Name Description Value
interimSchemaName Schema name for interim table. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
keys Key column names for unique row identification. Type: array of strings (or Expression with resultType array of strings). For Bicep, you can use the any() function.

SqlMISink

Name Description Value
type Copy sink type. 'SqlMISink' (required)
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlWriterStoredProcedureName SQL writer stored procedure name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlWriterTableType SQL writer table type. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlWriterUseTableLock Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
storedProcedureParameters SQL stored procedure parameters. For Bicep, you can use the any() function.
storedProcedureTableTypeParameterName The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
tableOption The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
upsertSettings SQL upsert settings. SqlUpsertSettings
writeBehavior White behavior when copying data into azure SQL MI. Type: string (or Expression with resultType string) For Bicep, you can use the any() function.

SqlServerSink

Name Description Value
type Copy sink type. 'SqlServerSink' (required)
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlWriterStoredProcedureName SQL writer stored procedure name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlWriterTableType SQL writer table type. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlWriterUseTableLock Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
storedProcedureParameters SQL stored procedure parameters. For Bicep, you can use the any() function.
storedProcedureTableTypeParameterName The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
tableOption The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
upsertSettings SQL upsert settings. SqlUpsertSettings
writeBehavior Write behavior when copying data into sql server. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SqlSink

Name Description Value
type Copy sink type. 'SqlSink' (required)
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlWriterStoredProcedureName SQL writer stored procedure name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlWriterTableType SQL writer table type. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlWriterUseTableLock Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
storedProcedureParameters SQL stored procedure parameters. For Bicep, you can use the any() function.
storedProcedureTableTypeParameterName The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
tableOption The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
upsertSettings SQL upsert settings. SqlUpsertSettings
writeBehavior Write behavior when copying data into sql. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

WarehouseSink

Name Description Value
type Copy sink type. 'WarehouseSink' (required)
allowCopyCommand Indicates to use Copy Command to copy data into SQL Data Warehouse. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
copyCommandSettings Specifies Copy Command related settings when allowCopyCommand is true. DWCopyCommandSettings
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
tableOption The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
writeBehavior Write behavior when copying data into azure Microsoft Fabric Data Warehouse. Type: DWWriteBehaviorEnum (or Expression with resultType DWWriteBehaviorEnum) For Bicep, you can use the any() function.

SkipErrorFile

Name Description Value
dataInconsistency Skip if source/sink file changed by other concurrent write. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
fileMissing Skip if file is deleted by other client during copy. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.

CopySource

Name Description Value
disableMetricsCollection If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
maxConcurrentConnections The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
sourceRetryCount Source retry count. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
sourceRetryWait Source retry wait. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
type Set the object type AmazonMWSSource
AmazonRdsForOracleSource
AmazonRdsForSqlServerSource
AmazonRedshiftSource
AvroSource
AzureBlobFSSource
AzureDatabricksDeltaLakeSource
AzureDataExplorerSource
AzureDataLakeStoreSource
AzureMariaDBSource
AzureMySqlSource
AzurePostgreSqlSource
AzureSqlSource
AzureTableSource
BinarySource
BlobSource
CassandraSource
CommonDataServiceForAppsSource
ConcurSource
CosmosDbMongoDbApiSource
CosmosDbSqlApiSource
CouchbaseSource
Db2Source
DelimitedTextSource
DocumentDbCollectionSource
DrillSource
DynamicsAXSource
DynamicsCrmSource
DynamicsSource
EloquaSource
ExcelSource
FileSystemSource
GoogleAdWordsSource
GoogleBigQuerySource
GoogleBigQueryV2Source
GreenplumSource
HBaseSource
HdfsSource
HiveSource
HttpSource
HubspotSource
ImpalaSource
InformixSource
JiraSource
JsonSource
LakeHouseTableSource
MagentoSource
MariaDBSource
MarketoSource
MicrosoftAccessSource
MongoDbAtlasSource
MongoDbSource
MongoDbV2Source
MySqlSource
NetezzaSource
ODataSource
OdbcSource
Office365Source
OracleServiceCloudSource
OracleSource
OrcSource
ParquetSource
PaypalSource
PhoenixSource
PostgreSqlSource
PostgreSqlV2Source
PrestoSource
QuickBooksSource
RelationalSource
ResponsysSource
RestSource
SalesforceMarketingCloudSource
SalesforceServiceCloudSource
SalesforceServiceCloudV2Source
SalesforceSource
SalesforceV2Source
SapBwSource
SapCloudForCustomerSource
SapEccSource
SapHanaSource
SapOdpSource
SapOpenHubSource
SapTableSource
ServiceNowSource
ServiceNowV2Source
SharePointOnlineListSource
ShopifySource
SnowflakeSource
SnowflakeV2Source
SparkSource
SqlDWSource
SqlMISource
SqlServerSource
SqlSource
SquareSource
SybaseSource
TeradataSource
VerticaSource
WarehouseSource
WebSource
XeroSource
XmlSource
ZohoSource (required)

AmazonMWSSource

Name Description Value
type Copy source type. 'AmazonMWSSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

AmazonRdsForOracleSource

Name Description Value
type Copy source type. 'AmazonRdsForOracleSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
oracleReaderQuery AmazonRdsForOracle reader query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionOption The partition mechanism that will be used for AmazonRdsForOracle read in parallel. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionSettings The settings that will be leveraged for AmazonRdsForOracle source partitioning. AmazonRdsForOraclePartitionSettings
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

AmazonRdsForOraclePartitionSettings

Name Description Value
partitionColumnName The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionLowerBound The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionNames Names of the physical partitions of AmazonRdsForOracle table. For Bicep, you can use the any() function.
partitionUpperBound The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

AmazonRdsForSqlServerSource

Name Description Value
type Copy source type. 'AmazonRdsForSqlServerSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
isolationLevel Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". For Bicep, you can use the any() function.
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
produceAdditionalTypes Which additional types to produce. For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
sqlReaderQuery SQL reader query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlReaderStoredProcedureName Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". For Bicep, you can use the any() function.

SqlPartitionSettings

Name Description Value
partitionColumnName The name of the column in integer or datetime type that will be used for proceeding partitioning. If not specified, the primary key of the table is auto-detected and used as the partition column. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionLowerBound The minimum value of the partition column for partition range splitting. This value is used to decide the partition stride, not for filtering the rows in table. All rows in the table or query result will be partitioned and copied. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionUpperBound The maximum value of the partition column for partition range splitting. This value is used to decide the partition stride, not for filtering the rows in table. All rows in the table or query result will be partitioned and copied. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

AmazonRedshiftSource

Name Description Value
type Copy source type. 'AmazonRedshiftSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query Database query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
redshiftUnloadSettings The Amazon S3 settings needed for the interim Amazon S3 when copying from Amazon Redshift with unload. With this, data from Amazon Redshift source will be unloaded into S3 first and then copied into the targeted sink from the interim S3. RedshiftUnloadSettings

RedshiftUnloadSettings

Name Description Value
bucketName The bucket of the interim Amazon S3 which will be used to store the unloaded data from Amazon Redshift source. The bucket must be in the same region as the Amazon Redshift source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
s3LinkedServiceName The name of the Amazon S3 linked service which will be used for the unload operation when copying from the Amazon Redshift source. LinkedServiceReference (required)

AvroSource

Name Description Value
type Copy source type. 'AvroSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
storeSettings Avro store settings. StoreReadSettings

StoreReadSettings

Name Description Value
disableMetricsCollection If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
maxConcurrentConnections The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
type Set the object type AmazonS3CompatibleReadSettings
AmazonS3ReadSettings
AzureBlobFSReadSettings
AzureBlobStorageReadSettings
AzureDataLakeStoreReadSettings
AzureFileStorageReadSettings
FileServerReadSettings
FtpReadSettings
GoogleCloudStorageReadSettings
HdfsReadSettings
HttpReadSettings
LakeHouseReadSettings
OracleCloudStorageReadSettings
SftpReadSettings (required)

AmazonS3CompatibleReadSettings

Name Description Value
type The read setting type. 'AmazonS3CompatibleReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
prefix The prefix filter for the S3 Compatible object name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
wildcardFileName Amazon S3 Compatible wildcardFileName. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
wildcardFolderPath Amazon S3 Compatible wildcardFolderPath. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

AmazonS3ReadSettings

Name Description Value
type The read setting type. 'AmazonS3ReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
prefix The prefix filter for the S3 object name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
wildcardFileName AmazonS3 wildcardFileName. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
wildcardFolderPath AmazonS3 wildcardFolderPath. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

AzureBlobFSReadSettings

Name Description Value
type The read setting type. 'AzureBlobFSReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
wildcardFileName Azure blobFS wildcardFileName. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
wildcardFolderPath Azure blobFS wildcardFolderPath. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

AzureBlobStorageReadSettings

Name Description Value
type The read setting type. 'AzureBlobStorageReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
prefix The prefix filter for the Azure Blob name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
wildcardFileName Azure blob wildcardFileName. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
wildcardFolderPath Azure blob wildcardFolderPath. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

AzureDataLakeStoreReadSettings

Name Description Value
type The read setting type. 'AzureDataLakeStoreReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
listAfter Lists files after the value (exclusive) based on file/folder names’ lexicographical order. Applies under the folderPath in data set, and filter files/sub-folders under the folderPath. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
listBefore Lists files before the value (inclusive) based on file/folder names’ lexicographical order. Applies under the folderPath in data set, and filter files/sub-folders under the folderPath. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
wildcardFileName ADLS wildcardFileName. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
wildcardFolderPath ADLS wildcardFolderPath. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

AzureFileStorageReadSettings

Name Description Value
type The read setting type. 'AzureFileStorageReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
prefix The prefix filter for the Azure File name starting from root path. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
wildcardFileName Azure File Storage wildcardFileName. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
wildcardFolderPath Azure File Storage wildcardFolderPath. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

FileServerReadSettings

Name Description Value
type The read setting type. 'FileServerReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
fileFilter Specify a filter to be used to select a subset of files in the folderPath rather than all files. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
wildcardFileName FileServer wildcardFileName. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
wildcardFolderPath FileServer wildcardFolderPath. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

FtpReadSettings

Name Description Value
type The read setting type. 'FtpReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
disableChunking If true, disable parallel reading within each file. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
useBinaryTransfer Specify whether to use binary transfer mode for FTP stores. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
wildcardFileName Ftp wildcardFileName. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
wildcardFolderPath Ftp wildcardFolderPath. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

GoogleCloudStorageReadSettings

Name Description Value
type The read setting type. 'GoogleCloudStorageReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
prefix The prefix filter for the Google Cloud Storage object name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
wildcardFileName Google Cloud Storage wildcardFileName. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
wildcardFolderPath Google Cloud Storage wildcardFolderPath. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

HdfsReadSettings

Name Description Value
type The read setting type. 'HdfsReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
distcpSettings Specifies Distcp-related settings. DistcpSettings
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
wildcardFileName HDFS wildcardFileName. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
wildcardFolderPath HDFS wildcardFolderPath. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

DistcpSettings

Name Description Value
distcpOptions Specifies the Distcp options. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
resourceManagerEndpoint Specifies the Yarn ResourceManager endpoint. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
tempScriptPath Specifies an existing folder path which will be used to store temp Distcp command script. The script file is generated by ADF and will be removed after Copy job finished. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)

HttpReadSettings

Name Description Value
type The read setting type. 'HttpReadSettings' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
additionalHeaders The additional HTTP headers in the request to the RESTful API. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
requestBody The HTTP request body to the RESTful API if requestMethod is POST. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
requestMethod The HTTP method used to call the RESTful API. The default is GET. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
requestTimeout Specifies the timeout for a HTTP client to get HTTP response from HTTP server. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

LakeHouseReadSettings

Name Description Value
type The read setting type. 'LakeHouseReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
wildcardFileName Microsoft Fabric LakeHouse Files wildcardFileName. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
wildcardFolderPath Microsoft Fabric LakeHouse Files wildcardFolderPath. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

OracleCloudStorageReadSettings

Name Description Value
type The read setting type. 'OracleCloudStorageReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
prefix The prefix filter for the Oracle Cloud Storage object name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
wildcardFileName Oracle Cloud Storage wildcardFileName. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
wildcardFolderPath Oracle Cloud Storage wildcardFolderPath. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SftpReadSettings

Name Description Value
type The read setting type. 'SftpReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
disableChunking If true, disable parallel reading within each file. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
wildcardFileName Sftp wildcardFileName. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
wildcardFolderPath Sftp wildcardFolderPath. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

AzureBlobFSSource

Name Description Value
type Copy source type. 'AzureBlobFSSource' (required)
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
skipHeaderLineCount Number of header lines to skip from each blob. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
treatEmptyAsNull Treat empty as null. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.

AzureDatabricksDeltaLakeSource

Name Description Value
type Copy source type. 'AzureDatabricksDeltaLakeSource' (required)
exportSettings Azure Databricks Delta Lake export settings. AzureDatabricksDeltaLakeExportCommand
query Azure Databricks Delta Lake Sql query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

AzureDatabricksDeltaLakeExportCommand

Name Description Value
dateFormat Specify the date format for the csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
timestampFormat Specify the timestamp format for the csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
type The export setting type. string (required)
{customized property} For Bicep, you can use the any() function.

AzureDataExplorerSource

Name Description Value
type Copy source type. 'AzureDataExplorerSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
noTruncation The name of the Boolean option that controls whether truncation is applied to result-sets that go beyond a certain row-count limit. For Bicep, you can use the any() function.
query Database query. Should be a Kusto Query Language (KQL) query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).. For Bicep, you can use the any() function.

AzureDataLakeStoreSource

Name Description Value
type Copy source type. 'AzureDataLakeStoreSource' (required)
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.

AzureMariaDBSource

Name Description Value
type Copy source type. 'AzureMariaDBSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

AzureMySqlSource

Name Description Value
type Copy source type. 'AzureMySqlSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query Database query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

AzurePostgreSqlSource

Name Description Value
type Copy source type. 'AzurePostgreSqlSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

AzureSqlSource

Name Description Value
type Copy source type. 'AzureSqlSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
isolationLevel Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
produceAdditionalTypes Which additional types to produce. For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
sqlReaderQuery SQL reader query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlReaderStoredProcedureName Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". For Bicep, you can use the any() function.

AzureTableSource

Name Description Value
type Copy source type. 'AzureTableSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
azureTableSourceIgnoreTableNotFound Azure Table source ignore table not found. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
azureTableSourceQuery Azure Table source query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

BinarySource

Name Description Value
type Copy source type. 'BinarySource' (required)
formatSettings Binary format settings. BinaryReadSettings
storeSettings Binary store settings. StoreReadSettings

BinaryReadSettings

Name Description Value
compressionProperties Compression settings. CompressionReadSettings
type The read setting type. string (required)
{customized property} For Bicep, you can use the any() function.

CompressionReadSettings

Name Description Value
type Set the object type TarGZipReadSettings
TarReadSettings
ZipDeflateReadSettings (required)

TarGZipReadSettings

Name Description Value
type The Compression setting type. 'TarGZipReadSettings' (required)
preserveCompressionFileNameAsFolder Preserve the compression file name as folder path. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.

TarReadSettings

Name Description Value
type The Compression setting type. 'TarReadSettings' (required)
preserveCompressionFileNameAsFolder Preserve the compression file name as folder path. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.

ZipDeflateReadSettings

Name Description Value
type The Compression setting type. 'ZipDeflateReadSettings' (required)
preserveZipFileNameAsFolder Preserve the zip file name as folder path. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.

BlobSource

Name Description Value
type Copy source type. 'BlobSource' (required)
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
skipHeaderLineCount Number of header lines to skip from each blob. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
treatEmptyAsNull Treat empty as null. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.

CassandraSource

Name Description Value
type Copy source type. 'CassandraSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
consistencyLevel The consistency level specifies how many Cassandra servers must respond to a read request before returning data to the client application. Cassandra checks the specified number of Cassandra servers for data to satisfy the read request. Must be one of cassandraSourceReadConsistencyLevels. The default value is 'ONE'. It is case-insensitive. 'ALL'
'EACH_QUORUM'
'LOCAL_ONE'
'LOCAL_QUORUM'
'LOCAL_SERIAL'
'ONE'
'QUORUM'
'SERIAL'
'THREE'
'TWO'
query Database query. Should be a SQL-92 query expression or Cassandra Query Language (CQL) command. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

CommonDataServiceForAppsSource

Name Description Value
type Copy source type. 'CommonDataServiceForAppsSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query FetchXML is a proprietary query language that is used in Microsoft Common Data Service for Apps (online & on-premises). Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

ConcurSource

Name Description Value
type Copy source type. 'ConcurSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

CosmosDbMongoDbApiSource

Name Description Value
type Copy source type. 'CosmosDbMongoDbApiSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
batchSize Specifies the number of documents to return in each batch of the response from MongoDB instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
cursorMethods Cursor methods for Mongodb query. MongoDbCursorMethodsProperties
filter Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

MongoDbCursorMethodsProperties

Name Description Value
limit Specifies the maximum number of documents the server returns. limit() is analogous to the LIMIT statement in a SQL database. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
project Specifies the fields to return in the documents that match the query filter. To return all fields in the matching documents, omit this parameter. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
skip Specifies the how many documents skipped and where MongoDB begins returning results. This approach may be useful in implementing paginated results. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
sort Specifies the order in which the query returns matching documents. Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
{customized property} For Bicep, you can use the any() function.

CosmosDbSqlApiSource

Name Description Value
type Copy source type. 'CosmosDbSqlApiSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
detectDatetime Whether detect primitive values as datetime values. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
pageSize Page size of the result. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
preferredRegions Preferred regions. Type: array of strings (or Expression with resultType array of strings). For Bicep, you can use the any() function.
query SQL API query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

CouchbaseSource

Name Description Value
type Copy source type. 'CouchbaseSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

Db2Source

Name Description Value
type Copy source type. 'Db2Source' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query Database query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

DelimitedTextSource

Name Description Value
type Copy source type. 'DelimitedTextSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
formatSettings DelimitedText format settings. DelimitedTextReadSettings
storeSettings DelimitedText store settings. StoreReadSettings

DelimitedTextReadSettings

Name Description Value
compressionProperties Compression settings. CompressionReadSettings
skipLineCount Indicates the number of non-empty rows to skip when reading data from input files. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
type The read setting type. string (required)
{customized property} For Bicep, you can use the any() function.

DocumentDbCollectionSource

Name Description Value
type Copy source type. 'DocumentDbCollectionSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
nestingSeparator Nested properties separator. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
query Documents query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

DrillSource

Name Description Value
type Copy source type. 'DrillSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

DynamicsAXSource

Name Description Value
type Copy source type. 'DynamicsAXSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

DynamicsCrmSource

Name Description Value
type Copy source type. 'DynamicsCrmSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query FetchXML is a proprietary query language that is used in Microsoft Dynamics CRM (online & on-premises). Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

DynamicsSource

Name Description Value
type Copy source type. 'DynamicsSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query FetchXML is a proprietary query language that is used in Microsoft Dynamics (online & on-premises). Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

EloquaSource

Name Description Value
type Copy source type. 'EloquaSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

ExcelSource

Name Description Value
type Copy source type. 'ExcelSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
storeSettings Excel store settings. StoreReadSettings

FileSystemSource

Name Description Value
type Copy source type. 'FileSystemSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.

GoogleAdWordsSource

Name Description Value
type Copy source type. 'GoogleAdWordsSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

GoogleBigQuerySource

Name Description Value
type Copy source type. 'GoogleBigQuerySource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

GoogleBigQueryV2Source

Name Description Value
type Copy source type. 'GoogleBigQueryV2Source' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

GreenplumSource

Name Description Value
type Copy source type. 'GreenplumSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

HBaseSource

Name Description Value
type Copy source type. 'HBaseSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

HdfsSource

Name Description Value
type Copy source type. 'HdfsSource' (required)
distcpSettings Specifies Distcp-related settings. DistcpSettings
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.

HiveSource

Name Description Value
type Copy source type. 'HiveSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

HttpSource

Name Description Value
type Copy source type. 'HttpSource' (required)
httpRequestTimeout Specifies the timeout for a HTTP client to get HTTP response from HTTP server. The default value is equivalent to System.Net.HttpWebRequest.Timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

HubspotSource

Name Description Value
type Copy source type. 'HubspotSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

ImpalaSource

Name Description Value
type Copy source type. 'ImpalaSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

InformixSource

Name Description Value
type Copy source type. 'InformixSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query Database query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

JiraSource

Name Description Value
type Copy source type. 'JiraSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

JsonSource

Name Description Value
type Copy source type. 'JsonSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
formatSettings Json format settings. JsonReadSettings
storeSettings Json store settings. StoreReadSettings

JsonReadSettings

Name Description Value
compressionProperties Compression settings. CompressionReadSettings
type The read setting type. string (required)
{customized property} For Bicep, you can use the any() function.

LakeHouseTableSource

Name Description Value
type Copy source type. 'LakeHouseTableSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
timestampAsOf Query an older snapshot by timestamp. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
versionAsOf Query an older snapshot by version. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.

MagentoSource

Name Description Value
type Copy source type. 'MagentoSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

MariaDBSource

Name Description Value
type Copy source type. 'MariaDBSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

MarketoSource

Name Description Value
type Copy source type. 'MarketoSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

MicrosoftAccessSource

Name Description Value
type Copy source type. 'MicrosoftAccessSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query Database query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

MongoDbAtlasSource

Name Description Value
type Copy source type. 'MongoDbAtlasSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
batchSize Specifies the number of documents to return in each batch of the response from MongoDB Atlas instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
cursorMethods Cursor methods for Mongodb query MongoDbCursorMethodsProperties
filter Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

MongoDbSource

Name Description Value
type Copy source type. 'MongoDbSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query Database query. Should be a SQL-92 query expression. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

MongoDbV2Source

Name Description Value
type Copy source type. 'MongoDbV2Source' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
batchSize Specifies the number of documents to return in each batch of the response from MongoDB instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
cursorMethods Cursor methods for Mongodb query MongoDbCursorMethodsProperties
filter Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

MySqlSource

Name Description Value
type Copy source type. 'MySqlSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query Database query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

NetezzaSource

Name Description Value
type Copy source type. 'NetezzaSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
partitionOption The partition mechanism that will be used for Netezza read in parallel. Possible values include: "None", "DataSlice", "DynamicRange". For Bicep, you can use the any() function.
partitionSettings The settings that will be leveraged for Netezza source partitioning. NetezzaPartitionSettings
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

NetezzaPartitionSettings

Name Description Value
partitionColumnName The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionLowerBound The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionUpperBound The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

ODataSource

Name Description Value
type Copy source type. 'ODataSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
query OData query. For example, "$top=1". Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

OdbcSource

Name Description Value
type Copy source type. 'OdbcSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query Database query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

Office365Source

Name Description Value
type Copy source type. 'Office365Source' (required)
allowedGroups The groups containing all the users. Type: array of strings (or Expression with resultType array of strings). For Bicep, you can use the any() function.
dateFilterColumn The Column to apply the {paramref name="StartTime"/} and {paramref name="EndTime"/}. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
endTime End time of the requested range for this dataset. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
outputColumns The columns to be read out from the Office 365 table. Type: array of objects (or Expression with resultType array of objects). itemType: OutputColumn. Example: [ { "name": "Id" }, { "name": "CreatedDateTime" } ] For Bicep, you can use the any() function.
startTime Start time of the requested range for this dataset. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
userScopeFilterUri The user scope uri. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

OracleServiceCloudSource

Name Description Value
type Copy source type. 'OracleServiceCloudSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

OracleSource

Name Description Value
type Copy source type. 'OracleSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
oracleReaderQuery Oracle reader query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionOption The partition mechanism that will be used for Oracle read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". For Bicep, you can use the any() function.
partitionSettings The settings that will be leveraged for Oracle source partitioning. OraclePartitionSettings
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

OraclePartitionSettings

Name Description Value
partitionColumnName The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionLowerBound The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionNames Names of the physical partitions of Oracle table. For Bicep, you can use the any() function.
partitionUpperBound The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

OrcSource

Name Description Value
type Copy source type. 'OrcSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
storeSettings ORC store settings. StoreReadSettings

ParquetSource

Name Description Value
type Copy source type. 'ParquetSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
formatSettings Parquet format settings. ParquetReadSettings
storeSettings Parquet store settings. StoreReadSettings

ParquetReadSettings

Name Description Value
compressionProperties Compression settings. CompressionReadSettings
type The read setting type. string (required)
{customized property} For Bicep, you can use the any() function.

PaypalSource

Name Description Value
type Copy source type. 'PaypalSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

PhoenixSource

Name Description Value
type Copy source type. 'PhoenixSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

PostgreSqlSource

Name Description Value
type Copy source type. 'PostgreSqlSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query Database query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

PostgreSqlV2Source

Name Description Value
type Copy source type. 'PostgreSqlV2Source' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query Database query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

PrestoSource

Name Description Value
type Copy source type. 'PrestoSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

QuickBooksSource

Name Description Value
type Copy source type. 'QuickBooksSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

RelationalSource

Name Description Value
type Copy source type. 'RelationalSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query Database query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

ResponsysSource

Name Description Value
type Copy source type. 'ResponsysSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

RestSource

Name Description Value
type Copy source type. 'RestSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: key value pairs (value should be string type). For Bicep, you can use the any() function.
additionalHeaders The additional HTTP headers in the request to the RESTful API. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
paginationRules The pagination rules to compose next page requests. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
requestBody The HTTP request body to the RESTful API if requestMethod is POST. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
requestInterval The time to await before sending next page request. For Bicep, you can use the any() function.
requestMethod The HTTP method used to call the RESTful API. The default is GET. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SalesforceMarketingCloudSource

Name Description Value
type Copy source type. 'SalesforceMarketingCloudSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

SalesforceServiceCloudSource

Name Description Value
type Copy source type. 'SalesforceServiceCloudSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query Database query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
readBehavior The read behavior for the operation. Default is Query. Allowed values: Query/QueryAll. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SalesforceServiceCloudV2Source

Name Description Value
type Copy source type. 'SalesforceServiceCloudV2Source' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
includeDeletedObjects This property control whether query result contains Deleted objects. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
SOQLQuery Database query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SalesforceSource

Name Description Value
type Copy source type. 'SalesforceSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query Database query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
readBehavior The read behavior for the operation. Default is Query. Allowed values: Query/QueryAll. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SalesforceV2Source

Name Description Value
type Copy source type. 'SalesforceV2Source' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
includeDeletedObjects This property control whether query result contains Deleted objects. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
SOQLQuery Database query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SapBwSource

Name Description Value
type Copy source type. 'SapBwSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query MDX query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

SapCloudForCustomerSource

Name Description Value
type Copy source type. 'SapCloudForCustomerSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
query SAP Cloud for Customer OData query. For example, "$top=1". Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

SapEccSource

Name Description Value
type Copy source type. 'SapEccSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
query SAP ECC OData query. For example, "$top=1". Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

SapHanaSource

Name Description Value
type Copy source type. 'SapHanaSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
packetSize The packet size of data read from SAP HANA. Type: integer(or Expression with resultType integer). For Bicep, you can use the any() function.
partitionOption The partition mechanism that will be used for SAP HANA read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "SapHanaDynamicRange". For Bicep, you can use the any() function.
partitionSettings The settings that will be leveraged for SAP HANA source partitioning. SapHanaPartitionSettings
query SAP HANA Sql query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

SapHanaPartitionSettings

Name Description Value
partitionColumnName The name of the column that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SapOdpSource

Name Description Value
type Copy source type. 'SapOdpSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
extractionMode The extraction mode. Allowed value include: Full, Delta and Recovery. The default value is Full. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
projection Specifies the columns to be selected from source data. Type: array of objects(projection) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
selection Specifies the selection conditions from source data. Type: array of objects(selection) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
subscriberProcess The subscriber process to manage the delta process. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SapOpenHubSource

Name Description Value
type Copy source type. 'SapOpenHubSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
baseRequestId The ID of request for delta loading. Once it is set, only data with requestId larger than the value of this property will be retrieved. The default value is 0. Type: integer (or Expression with resultType integer ). For Bicep, you can use the any() function.
customRfcReadTableFunctionModule Specifies the custom RFC function module that will be used to read data from SAP Table. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
excludeLastRequest Whether to exclude the records of the last request. The default value is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
sapDataColumnDelimiter The single character that will be used as delimiter passed to SAP RFC as well as splitting the output data retrieved. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SapTableSource

Name Description Value
type Copy source type. 'SapTableSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
batchSize Specifies the maximum number of rows that will be retrieved at a time when retrieving data from SAP Table. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
customRfcReadTableFunctionModule Specifies the custom RFC function module that will be used to read data from SAP Table. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionOption The partition mechanism that will be used for SAP table read in parallel. Possible values include: "None", "PartitionOnInt", "PartitionOnCalendarYear", "PartitionOnCalendarMonth", "PartitionOnCalendarDate", "PartitionOnTime". For Bicep, you can use the any() function.
partitionSettings The settings that will be leveraged for SAP table source partitioning. SapTablePartitionSettings
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
rfcTableFields The fields of the SAP table that will be retrieved. For example, column0, column1. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
rfcTableOptions The options for the filtering of the SAP Table. For example, COLUMN0 EQ SOME VALUE. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
rowCount The number of rows to be retrieved. Type: integer(or Expression with resultType integer). For Bicep, you can use the any() function.
rowSkips The number of rows that will be skipped. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
sapDataColumnDelimiter The single character that will be used as delimiter passed to SAP RFC as well as splitting the output data retrieved. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SapTablePartitionSettings

Name Description Value
maxPartitionsNumber The maximum value of partitions the table will be split into. Type: integer (or Expression with resultType string). For Bicep, you can use the any() function.
partitionColumnName The name of the column that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionLowerBound The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionUpperBound The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

ServiceNowSource

Name Description Value
type Copy source type. 'ServiceNowSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

ServiceNowV2Source

Name Description Value
type Copy source type. 'ServiceNowV2Source' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
expression Expression to filter data from source. ExpressionV2
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

ExpressionV2

Name Description Value
operands List of nested expressions. ExpressionV2[]
operator Expression operator value Type: string. string
type Type of expressions supported by the system. Type: string. 'Binary'
'Constant'
'Field'
'Unary'
value Value for Constant/Field Type: string. string

SharePointOnlineListSource

Name Description Value
type Copy source type. 'SharePointOnlineListSource' (required)
httpRequestTimeout The wait time to get a response from SharePoint Online. Default value is 5 minutes (00:05:00). Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
query The OData query to filter the data in SharePoint Online list. For example, "$top=1". Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

ShopifySource

Name Description Value
type Copy source type. 'ShopifySource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

SnowflakeSource

Name Description Value
type Copy source type. 'SnowflakeSource' (required)
exportSettings Snowflake export settings. SnowflakeExportCopyCommand (required)
query Snowflake Sql query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SnowflakeExportCopyCommand

Name Description Value
additionalCopyOptions Additional copy options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": "'HH24:MI:SS.FF'" } SnowflakeExportCopyCommandAdditionalCopyOptions
additionalFormatOptions Additional format options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalFormatOptions": { "OVERWRITE": "TRUE", "MAX_FILE_SIZE": "'FALSE'" } SnowflakeExportCopyCommandAdditionalFormatOptions
type The export setting type. string (required)
{customized property} For Bicep, you can use the any() function.

SnowflakeExportCopyCommandAdditionalCopyOptions

Name Description Value
{customized property} For Bicep, you can use the any() function.

SnowflakeExportCopyCommandAdditionalFormatOptions

Name Description Value
{customized property} For Bicep, you can use the any() function.

SnowflakeV2Source

Name Description Value
type Copy source type. 'SnowflakeV2Source' (required)
exportSettings Snowflake export settings. SnowflakeExportCopyCommand (required)
query Snowflake Sql query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SparkSource

Name Description Value
type Copy source type. 'SparkSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

SqlDWSource

Name Description Value
type Copy source type. 'SqlDWSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
isolationLevel Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
sqlReaderQuery SQL Data Warehouse reader query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlReaderStoredProcedureName Name of the stored procedure for a SQL Data Warehouse source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". Type: object (or Expression with resultType object), itemType: StoredProcedureParameter. For Bicep, you can use the any() function.

SqlMISource

Name Description Value
type Copy source type. 'SqlMISource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
isolationLevel Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
produceAdditionalTypes Which additional types to produce. For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
sqlReaderQuery SQL reader query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlReaderStoredProcedureName Name of the stored procedure for a Azure SQL Managed Instance source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". For Bicep, you can use the any() function.

SqlServerSource

Name Description Value
type Copy source type. 'SqlServerSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
isolationLevel Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
produceAdditionalTypes Which additional types to produce. For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
sqlReaderQuery SQL reader query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlReaderStoredProcedureName Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". For Bicep, you can use the any() function.

SqlSource

Name Description Value
type Copy source type. 'SqlSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
isolationLevel Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
sqlReaderQuery SQL reader query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlReaderStoredProcedureName Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". For Bicep, you can use the any() function.

SquareSource

Name Description Value
type Copy source type. 'SquareSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

SybaseSource

Name Description Value
type Copy source type. 'SybaseSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query Database query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

TeradataSource

Name Description Value
type Copy source type. 'TeradataSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
partitionOption The partition mechanism that will be used for teradata read in parallel. Possible values include: "None", "Hash", "DynamicRange". For Bicep, you can use the any() function.
partitionSettings The settings that will be leveraged for teradata source partitioning. TeradataPartitionSettings
query Teradata query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

TeradataPartitionSettings

Name Description Value
partitionColumnName The name of the column that will be used for proceeding range or hash partitioning. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionLowerBound The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionUpperBound The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

VerticaSource

Name Description Value
type Copy source type. 'VerticaSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

WarehouseSource

Name Description Value
type Copy source type. 'WarehouseSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
isolationLevel Specifies the transaction locking behavior for the Microsoft Fabric Warehouse source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". For Bicep, you can use the any() function.
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
sqlReaderQuery Microsoft Fabric Warehouse reader query. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
sqlReaderStoredProcedureName Name of the stored procedure for a Microsoft Fabric Warehouse source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". Type: object (or Expression with resultType object), itemType: StoredProcedureParameter. For Bicep, you can use the any() function.

WebSource

Name Description Value
type Copy source type. 'WebSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.

XeroSource

Name Description Value
type Copy source type. 'XeroSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

XmlSource

Name Description Value
type Copy source type. 'XmlSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
formatSettings Xml format settings. XmlReadSettings
storeSettings Xml store settings. StoreReadSettings

XmlReadSettings

Name Description Value
compressionProperties Compression settings. CompressionReadSettings
detectDataType Indicates whether type detection is enabled when reading the xml files. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
namespacePrefixes Namespace uri to prefix mappings to override the prefixes in column names when namespace is enabled, if no prefix is defined for a namespace uri, the prefix of xml element/attribute name in the xml data file will be used. Example: "{"http://www.example.com/xml":"prefix"}" Type: object (or Expression with resultType object). For Bicep, you can use the any() function.
namespaces Indicates whether namespace is enabled when reading the xml files. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
type The read setting type. string (required)
validationMode Indicates what validation method is used when reading the xml files. Allowed values: 'none', 'xsd', or 'dtd'. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
{customized property} For Bicep, you can use the any() function.

ZohoSource

Name Description Value
type Copy source type. 'ZohoSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). For Bicep, you can use the any() function.
query A query to retrieve data from source. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

StagingSettings

Name Description Value
enableCompression Specifies whether to use compression when copying data via an interim staging. Default value is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
linkedServiceName Staging linked service reference. LinkedServiceReference (required)
path The path to storage for storing the interim data. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
{customized property} For Bicep, you can use the any() function.

CustomActivity

Name Description Value
type Type of activity. 'Custom' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Custom activity properties. CustomActivityTypeProperties (required)

CustomActivityTypeProperties

Name Description Value
autoUserSpecification Elevation level and scope for the user, default is nonadmin task. Type: string (or Expression with resultType double). For Bicep, you can use the any() function.
command Command for custom activity Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
extendedProperties User defined property bag. There is no restriction on the keys or values that can be used. The user specified custom activity has the full responsibility to consume and interpret the content defined. CustomActivityTypePropertiesExtendedProperties
folderPath Folder path for resource files Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
referenceObjects Reference objects CustomActivityReferenceObject
resourceLinkedService Resource linked service reference. LinkedServiceReference
retentionTimeInDays The retention time for the files submitted for custom activity. Type: double (or Expression with resultType double). For Bicep, you can use the any() function.

CustomActivityTypePropertiesExtendedProperties

Name Description Value
{customized property} For Bicep, you can use the any() function.

CustomActivityReferenceObject

Name Description Value
datasets Dataset references. DatasetReference[]
linkedServices Linked service references. LinkedServiceReference[]

DatabricksNotebookActivity

Name Description Value
type Type of activity. 'DatabricksNotebook' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Databricks Notebook activity properties. DatabricksNotebookActivityTypeProperties (required)

DatabricksNotebookActivityTypeProperties

Name Description Value
baseParameters Base parameters to be used for each run of this job.If the notebook takes a parameter that is not specified, the default value from the notebook will be used. DatabricksNotebookActivityTypePropertiesBaseParamete...
libraries A list of libraries to be installed on the cluster that will execute the job. DatabricksNotebookActivityTypePropertiesLibrariesIte...[]
notebookPath The absolute path of the notebook to be run in the Databricks Workspace. This path must begin with a slash. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)

DatabricksNotebookActivityTypePropertiesBaseParamete...

Name Description Value
{customized property} For Bicep, you can use the any() function.

DatabricksNotebookActivityTypePropertiesLibrariesIte...

Name Description Value
{customized property} For Bicep, you can use the any() function.

DatabricksSparkJarActivity

Name Description Value
type Type of activity. 'DatabricksSparkJar' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Databricks SparkJar activity properties. DatabricksSparkJarActivityTypeProperties (required)

DatabricksSparkJarActivityTypeProperties

Name Description Value
libraries A list of libraries to be installed on the cluster that will execute the job. DatabricksSparkJarActivityTypePropertiesLibrariesIte...[]
mainClassName The full name of the class containing the main method to be executed. This class must be contained in a JAR provided as a library. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
parameters Parameters that will be passed to the main method. any[]

DatabricksSparkJarActivityTypePropertiesLibrariesIte...

Name Description Value
{customized property} For Bicep, you can use the any() function.

DatabricksSparkPythonActivity

Name Description Value
type Type of activity. 'DatabricksSparkPython' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Databricks SparkPython activity properties. DatabricksSparkPythonActivityTypeProperties (required)

DatabricksSparkPythonActivityTypeProperties

Name Description Value
libraries A list of libraries to be installed on the cluster that will execute the job. DatabricksSparkPythonActivityTypePropertiesLibraries...[]
parameters Command line parameters that will be passed to the Python file. any[]
pythonFile The URI of the Python file to be executed. DBFS paths are supported. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)

DatabricksSparkPythonActivityTypePropertiesLibraries...

Name Description Value
{customized property} For Bicep, you can use the any() function.

DataLakeAnalyticsUsqlActivity

Name Description Value
type Type of activity. 'DataLakeAnalyticsU-SQL' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Data Lake Analytics U-SQL activity properties. DataLakeAnalyticsUsqlActivityTypeProperties (required)

DataLakeAnalyticsUsqlActivityTypeProperties

Name Description Value
compilationMode Compilation mode of U-SQL. Must be one of these values : Semantic, Full and SingleBox. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
degreeOfParallelism The maximum number of nodes simultaneously used to run the job. Default value is 1. Type: integer (or Expression with resultType integer), minimum: 1. For Bicep, you can use the any() function.
parameters Parameters for U-SQL job request. DataLakeAnalyticsUsqlActivityTypePropertiesParameter...
priority Determines which jobs out of all that are queued should be selected to run first. The lower the number, the higher the priority. Default value is 1000. Type: integer (or Expression with resultType integer), minimum: 1. For Bicep, you can use the any() function.
runtimeVersion Runtime version of the U-SQL engine to use. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
scriptLinkedService Script linked service reference. LinkedServiceReference (required)
scriptPath Case-sensitive path to folder that contains the U-SQL script. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)

DataLakeAnalyticsUsqlActivityTypePropertiesParameter...

Name Description Value
{customized property} For Bicep, you can use the any() function.

DeleteActivity

Name Description Value
type Type of activity. 'Delete' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Delete activity properties. DeleteActivityTypeProperties (required)

DeleteActivityTypeProperties

Name Description Value
dataset Delete activity dataset reference. DatasetReference (required)
enableLogging Whether to record detailed logs of delete-activity execution. Default value is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
logStorageSettings Log storage settings customer need to provide when enableLogging is true. LogStorageSettings
maxConcurrentConnections The max concurrent connections to connect data source at the same time. int

Constraints:
Min value = 1
recursive If true, files or sub-folders under current folder path will be deleted recursively. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
storeSettings Delete activity store settings. StoreReadSettings

ExecuteDataFlowActivity

Name Description Value
type Type of activity. 'ExecuteDataFlow' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Execute data flow activity properties. ExecuteDataFlowActivityTypeProperties (required)

ExecuteDataFlowActivityTypeProperties

Name Description Value
compute Compute properties for data flow activity. ExecuteDataFlowActivityTypePropertiesCompute
continueOnError Continue on error setting used for data flow execution. Enables processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean) For Bicep, you can use the any() function.
dataFlow Data flow reference. DataFlowReference (required)
integrationRuntime The integration runtime reference. IntegrationRuntimeReference
runConcurrently Concurrent run setting used for data flow execution. Allows sinks with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean) For Bicep, you can use the any() function.
sourceStagingConcurrency Specify number of parallel staging for sources applicable to the sink. Type: integer (or Expression with resultType integer) For Bicep, you can use the any() function.
staging Staging info for execute data flow activity. DataFlowStagingInfo
traceLevel Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string) For Bicep, you can use the any() function.

ExecuteDataFlowActivityTypePropertiesCompute

Name Description Value
computeType Compute type of the cluster which will execute data flow job. Possible values include: 'General', 'MemoryOptimized', 'ComputeOptimized'. Type: string (or Expression with resultType string) For Bicep, you can use the any() function.
coreCount Core count of the cluster which will execute data flow job. Supported values are: 8, 16, 32, 48, 80, 144 and 272. Type: integer (or Expression with resultType integer) For Bicep, you can use the any() function.

DataFlowReference

Name Description Value
datasetParameters Reference data flow parameters from dataset. For Bicep, you can use the any() function.
parameters Data flow parameters ParameterValueSpecification
referenceName Reference data flow name. string (required)
type Data flow reference type. 'DataFlowReference' (required)
{customized property} For Bicep, you can use the any() function.

IntegrationRuntimeReference

Name Description Value
parameters Arguments for integration runtime. ParameterValueSpecification
referenceName Reference integration runtime name. string (required)
type Type of integration runtime. 'IntegrationRuntimeReference' (required)

DataFlowStagingInfo

Name Description Value
folderPath Folder path for staging blob. Type: string (or Expression with resultType string) For Bicep, you can use the any() function.
linkedService Staging linked service reference. LinkedServiceReference

ExecutePipelineActivity

Name Description Value
type Type of activity. 'ExecutePipeline' (required)
policy Execute pipeline activity policy. ExecutePipelineActivityPolicy
typeProperties Execute pipeline activity properties. ExecutePipelineActivityTypeProperties (required)

ExecutePipelineActivityPolicy

Name Description Value
secureInput When set to true, Input from activity is considered as secure and will not be logged to monitoring. bool
{customized property} For Bicep, you can use the any() function.

ExecutePipelineActivityTypeProperties

Name Description Value
parameters Pipeline parameters. ParameterValueSpecification
pipeline Pipeline reference. PipelineReference (required)
waitOnCompletion Defines whether activity execution will wait for the dependent pipeline execution to finish. Default is false. bool

PipelineReference

Name Description Value
name Reference name. string
referenceName Reference pipeline name. string (required)
type Pipeline reference type. 'PipelineReference' (required)

ExecuteSsisPackageActivity

Name Description Value
type Type of activity. 'ExecuteSSISPackage' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Execute SSIS package activity properties. ExecuteSsisPackageActivityTypeProperties (required)

ExecuteSsisPackageActivityTypeProperties

Name Description Value
connectVia The integration runtime reference. IntegrationRuntimeReference (required)
environmentPath The environment path to execute the SSIS package. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
executionCredential The package execution credential. SsisExecutionCredential
loggingLevel The logging level of SSIS package execution. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
logLocation SSIS package execution log location. SsisLogLocation
packageConnectionManagers The package level connection managers to execute the SSIS package. ExecuteSsisPackageActivityTypePropertiesPackageConne...
packageLocation SSIS package location. SsisPackageLocation (required)
packageParameters The package level parameters to execute the SSIS package. ExecuteSsisPackageActivityTypePropertiesPackageParam...
projectConnectionManagers The project level connection managers to execute the SSIS package. ExecuteSsisPackageActivityTypePropertiesProjectConne...
projectParameters The project level parameters to execute the SSIS package. ExecuteSsisPackageActivityTypePropertiesProjectParam...
propertyOverrides The property overrides to execute the SSIS package. ExecuteSsisPackageActivityTypePropertiesPropertyOver...
runtime Specifies the runtime to execute SSIS package. The value should be "x86" or "x64". Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SsisExecutionCredential

Name Description Value
domain Domain for windows authentication. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
password Password for windows authentication. SecureString (required)
userName UseName for windows authentication. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)

SecureString

Name Description Value
type Type of the secret. string (required)
value Value of secure string. string (required)

SsisLogLocation

Name Description Value
logPath The SSIS package execution log path. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
type The type of SSIS log location. 'File' (required)
typeProperties SSIS package execution log location properties. SsisLogLocationTypeProperties (required)

SsisLogLocationTypeProperties

Name Description Value
accessCredential The package execution log access credential. SsisAccessCredential
logRefreshInterval Specifies the interval to refresh log. The default interval is 5 minutes. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

SsisAccessCredential

Name Description Value
domain Domain for windows authentication. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
password Password for windows authentication. SecretBase (required)
userName UseName for windows authentication. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)

SecretBase

Name Description Value
type Set the object type AzureKeyVaultSecret
SecureString (required)

AzureKeyVaultSecretReference

Name Description Value
secretName The name of the secret in Azure Key Vault. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
secretVersion The version of the secret in Azure Key Vault. The default value is the latest version of the secret. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
store The Azure Key Vault linked service reference. LinkedServiceReference (required)
type Type of the secret. string (required)

ExecuteSsisPackageActivityTypePropertiesPackageConne...

Name Description Value
{customized property} SsisConnectionManager

SsisConnectionManager

Name Description Value
{customized property} SsisExecutionParameter

SsisExecutionParameter

Name Description Value
value SSIS package execution parameter value. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)

SsisPackageLocation

Name Description Value
packagePath The SSIS package path. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
type The type of SSIS package location. 'File'
'InlinePackage'
'PackageStore'
'SSISDB'
typeProperties SSIS package location properties. SsisPackageLocationTypeProperties

SsisPackageLocationTypeProperties

Name Description Value
accessCredential The package access credential. SsisAccessCredential
childPackages The embedded child package list. SsisChildPackage[]
configurationAccessCredential The configuration file access credential. SsisAccessCredential
configurationPath The configuration file of the package execution. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
packageContent The embedded package content. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
packageLastModifiedDate The embedded package last modified date. string
packageName The package name. string
packagePassword Password of the package. SecretBase

SsisChildPackage

Name Description Value
packageContent Content for embedded child package. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
packageLastModifiedDate Last modified date for embedded child package. string
packageName Name for embedded child package. string
packagePath Path for embedded child package. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)

ExecuteSsisPackageActivityTypePropertiesPackageParam...

Name Description Value
{customized property} SsisExecutionParameter

ExecuteSsisPackageActivityTypePropertiesProjectConne...

Name Description Value
{customized property} SsisConnectionManager

ExecuteSsisPackageActivityTypePropertiesProjectParam...

Name Description Value
{customized property} SsisExecutionParameter

ExecuteSsisPackageActivityTypePropertiesPropertyOver...

Name Description Value
{customized property} SsisPropertyOverride

SsisPropertyOverride

Name Description Value
isSensitive Whether SSIS package property override value is sensitive data. Value will be encrypted in SSISDB if it is true bool
value SSIS package property override value. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)

ExecuteWranglingDataflowActivity

Name Description Value
type Type of activity. 'ExecuteWranglingDataflow' (required)
policy Activity policy. ActivityPolicy
typeProperties Execute power query activity properties. ExecutePowerQueryActivityTypeProperties (required)

ExecutePowerQueryActivityTypeProperties

Name Description Value
compute Compute properties for data flow activity. ExecuteDataFlowActivityTypePropertiesCompute
continueOnError Continue on error setting used for data flow execution. Enables processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean) For Bicep, you can use the any() function.
dataFlow Data flow reference. DataFlowReference (required)
integrationRuntime The integration runtime reference. IntegrationRuntimeReference
queries List of mapping for Power Query mashup query to sink dataset(s). PowerQuerySinkMapping[]
runConcurrently Concurrent run setting used for data flow execution. Allows sinks with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean) For Bicep, you can use the any() function.
sinks (Deprecated. Please use Queries). List of Power Query activity sinks mapped to a queryName. ExecutePowerQueryActivityTypePropertiesSinks
sourceStagingConcurrency Specify number of parallel staging for sources applicable to the sink. Type: integer (or Expression with resultType integer) For Bicep, you can use the any() function.
staging Staging info for execute data flow activity. DataFlowStagingInfo
traceLevel Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string) For Bicep, you can use the any() function.

PowerQuerySinkMapping

Name Description Value
dataflowSinks List of sinks mapped to Power Query mashup query. PowerQuerySink[]
queryName Name of the query in Power Query mashup document. string

PowerQuerySink

Name Description Value
dataset Dataset reference. DatasetReference
description Transformation description. string
flowlet Flowlet Reference DataFlowReference
linkedService Linked service reference. LinkedServiceReference
name Transformation name. string (required)
rejectedDataLinkedService Rejected data linked service reference. LinkedServiceReference
schemaLinkedService Schema linked service reference. LinkedServiceReference
script sink script. string

ExecutePowerQueryActivityTypePropertiesSinks

Name Description Value
{customized property} PowerQuerySink

FailActivity

Name Description Value
type Type of activity. 'Fail' (required)
typeProperties Fail activity properties. FailActivityTypeProperties (required)

FailActivityTypeProperties

Name Description Value
errorCode The error code that categorizes the error type of the Fail activity. It can be dynamic content that's evaluated to a non empty/blank string at runtime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
message The error message that surfaced in the Fail activity. It can be dynamic content that's evaluated to a non empty/blank string at runtime. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)

FilterActivity

Name Description Value
type Type of activity. 'Filter' (required)
typeProperties Filter activity properties. FilterActivityTypeProperties (required)

FilterActivityTypeProperties

Name Description Value
condition Condition to be used for filtering the input. Expression (required)
items Input array on which filter should be applied. Expression (required)

Expression

Name Description Value
type Expression type. 'Expression' (required)
value Expression value. string (required)

ForEachActivity

Name Description Value
type Type of activity. 'ForEach' (required)
typeProperties ForEach activity properties. ForEachActivityTypeProperties (required)

ForEachActivityTypeProperties

Name Description Value
activities List of activities to execute . Activity[] (required)
batchCount Batch count to be used for controlling the number of parallel execution (when isSequential is set to false). int

Constraints:
Max value = 50
isSequential Should the loop be executed in sequence or in parallel (max 50) bool
items Collection to iterate. Expression (required)

GetMetadataActivity

Name Description Value
type Type of activity. 'GetMetadata' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties GetMetadata activity properties. GetMetadataActivityTypeProperties (required)

GetMetadataActivityTypeProperties

Name Description Value
dataset GetMetadata activity dataset reference. DatasetReference (required)
fieldList Fields of metadata to get from dataset. any[]
formatSettings GetMetadata activity format settings. FormatReadSettings
storeSettings GetMetadata activity store settings. StoreReadSettings

FormatReadSettings

Name Description Value
type Set the object type BinaryReadSettings
DelimitedTextReadSettings
JsonReadSettings
ParquetReadSettings
XmlReadSettings (required)

HDInsightHiveActivity

Name Description Value
type Type of activity. 'HDInsightHive' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties HDInsight Hive activity properties. HDInsightHiveActivityTypeProperties (required)

HDInsightHiveActivityTypeProperties

Name Description Value
arguments User specified arguments to HDInsightActivity. any[]
defines Allows user to specify defines for Hive job request. HDInsightHiveActivityTypePropertiesDefines
getDebugInfo Debug info option. 'Always'
'Failure'
'None'
queryTimeout Query timeout value (in minutes). Effective when the HDInsight cluster is with ESP (Enterprise Security Package) int
scriptLinkedService Script linked service reference. LinkedServiceReference
scriptPath Script path. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
storageLinkedServices Storage linked service references. LinkedServiceReference[]
variables User specified arguments under hivevar namespace. HDInsightHiveActivityTypePropertiesVariables

HDInsightHiveActivityTypePropertiesDefines

Name Description Value
{customized property} For Bicep, you can use the any() function.

HDInsightHiveActivityTypePropertiesVariables

Name Description Value
{customized property} For Bicep, you can use the any() function.

HDInsightMapReduceActivity

Name Description Value
type Type of activity. 'HDInsightMapReduce' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties HDInsight MapReduce activity properties. HDInsightMapReduceActivityTypeProperties (required)

HDInsightMapReduceActivityTypeProperties

Name Description Value
arguments User specified arguments to HDInsightActivity. any[]
className Class name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
defines Allows user to specify defines for the MapReduce job request. HDInsightMapReduceActivityTypePropertiesDefines
getDebugInfo Debug info option. 'Always'
'Failure'
'None'
jarFilePath Jar path. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
jarLibs Jar libs. any[]
jarLinkedService Jar linked service reference. LinkedServiceReference
storageLinkedServices Storage linked service references. LinkedServiceReference[]

HDInsightMapReduceActivityTypePropertiesDefines

Name Description Value
{customized property} For Bicep, you can use the any() function.

HDInsightPigActivity

Name Description Value
type Type of activity. 'HDInsightPig' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties HDInsight Pig activity properties. HDInsightPigActivityTypeProperties (required)

HDInsightPigActivityTypeProperties

Name Description Value
arguments User specified arguments to HDInsightActivity. Type: array (or Expression with resultType array). For Bicep, you can use the any() function.
defines Allows user to specify defines for Pig job request. HDInsightPigActivityTypePropertiesDefines
getDebugInfo Debug info option. 'Always'
'Failure'
'None'
scriptLinkedService Script linked service reference. LinkedServiceReference
scriptPath Script path. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
storageLinkedServices Storage linked service references. LinkedServiceReference[]

HDInsightPigActivityTypePropertiesDefines

Name Description Value
{customized property} For Bicep, you can use the any() function.

HDInsightSparkActivity

Name Description Value
type Type of activity. 'HDInsightSpark' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties HDInsight spark activity properties. HDInsightSparkActivityTypeProperties (required)

HDInsightSparkActivityTypeProperties

Name Description Value
arguments The user-specified arguments to HDInsightSparkActivity. any[]
className The application's Java/Spark main class. string
entryFilePath The relative path to the root folder of the code/package to be executed. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
getDebugInfo Debug info option. 'Always'
'Failure'
'None'
proxyUser The user to impersonate that will execute the job. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
rootPath The root path in 'sparkJobLinkedService' for all the job’s files. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
sparkConfig Spark configuration property. HDInsightSparkActivityTypePropertiesSparkConfig
sparkJobLinkedService The storage linked service for uploading the entry file and dependencies, and for receiving logs. LinkedServiceReference

HDInsightSparkActivityTypePropertiesSparkConfig

Name Description Value
{customized property} For Bicep, you can use the any() function.

HDInsightStreamingActivity

Name Description Value
type Type of activity. 'HDInsightStreaming' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties HDInsight streaming activity properties. HDInsightStreamingActivityTypeProperties (required)

HDInsightStreamingActivityTypeProperties

Name Description Value
arguments User specified arguments to HDInsightActivity. any[]
combiner Combiner executable name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
commandEnvironment Command line environment values. any[]
defines Allows user to specify defines for streaming job request. HDInsightStreamingActivityTypePropertiesDefines
fileLinkedService Linked service reference where the files are located. LinkedServiceReference
filePaths Paths to streaming job files. Can be directories. any[] (required)
getDebugInfo Debug info option. 'Always'
'Failure'
'None'
input Input blob path. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
mapper Mapper executable name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
output Output blob path. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
reducer Reducer executable name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
storageLinkedServices Storage linked service references. LinkedServiceReference[]

HDInsightStreamingActivityTypePropertiesDefines

Name Description Value
{customized property} For Bicep, you can use the any() function.

IfConditionActivity

Name Description Value
type Type of activity. 'IfCondition' (required)
typeProperties IfCondition activity properties. IfConditionActivityTypeProperties (required)

IfConditionActivityTypeProperties

Name Description Value
expression An expression that would evaluate to Boolean. This is used to determine the block of activities (ifTrueActivities or ifFalseActivities) that will be executed. Expression (required)
ifFalseActivities List of activities to execute if expression is evaluated to false. This is an optional property and if not provided, the activity will exit without any action. Activity[]
ifTrueActivities List of activities to execute if expression is evaluated to true. This is an optional property and if not provided, the activity will exit without any action. Activity[]

LookupActivity

Name Description Value
type Type of activity. 'Lookup' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Lookup activity properties. LookupActivityTypeProperties (required)

LookupActivityTypeProperties

Name Description Value
dataset Lookup activity dataset reference. DatasetReference (required)
firstRowOnly Whether to return first row or all rows. Default value is true. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
source Dataset-specific source properties, same as copy activity source. CopySource (required)

ScriptActivity

Name Description Value
type Type of activity. 'Script' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Script activity properties. ScriptActivityTypeProperties (required)

ScriptActivityTypeProperties

Name Description Value
logSettings Log settings of script activity. ScriptActivityTypePropertiesLogSettings
scriptBlockExecutionTimeout ScriptBlock execution timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.
scripts Array of script blocks. Type: array. ScriptActivityScriptBlock[]

ScriptActivityTypePropertiesLogSettings

Name Description Value
logDestination The destination of logs. Type: string. 'ActivityOutput'
'ExternalStore' (required)
logLocationSettings Log location settings customer needs to provide when enabling log. LogLocationSettings

ScriptActivityScriptBlock

Name Description Value
parameters Array of script parameters. Type: array. ScriptActivityParameter[]
text The query text. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
type The type of the query. Type: string. 'NonQuery'
'Query' (required)

ScriptActivityParameter

Name Description Value
direction The direction of the parameter. 'Input'
'InputOutput'
'Output'
name The name of the parameter. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
size The size of the output direction parameter. int
type The type of the parameter. 'Boolean'
'DateTime'
'DateTimeOffset'
'Decimal'
'Double'
'Guid'
'Int16'
'Int32'
'Int64'
'Single'
'String'
'Timespan'
value The value of the parameter. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SetVariableActivity

Name Description Value
type Type of activity. 'SetVariable' (required)
policy Activity policy. SecureInputOutputPolicy
typeProperties Set Variable activity properties. SetVariableActivityTypeProperties (required)

SecureInputOutputPolicy

Name Description Value
secureInput When set to true, Input from activity is considered as secure and will not be logged to monitoring. bool
secureOutput When set to true, Output from activity is considered as secure and will not be logged to monitoring. bool

SetVariableActivityTypeProperties

Name Description Value
setSystemVariable If set to true, it sets the pipeline run return value. bool
value Value to be set. Could be a static value or Expression. For Bicep, you can use the any() function.
variableName Name of the variable whose value needs to be set. string

SynapseSparkJobDefinitionActivity

Name Description Value
type Type of activity. 'SparkJob' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Execute spark job activity properties. SynapseSparkJobActivityTypeProperties (required)

SynapseSparkJobActivityTypeProperties

Name Description Value
args User specified arguments to SynapseSparkJobDefinitionActivity. any[]
className The fully-qualified identifier or the main class that is in the main definition file, which will override the 'className' of the spark job definition you provide. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
conf Spark configuration properties, which will override the 'conf' of the spark job definition you provide. For Bicep, you can use the any() function.
configurationType The type of the spark config. 'Artifact'
'Customized'
'Default'
driverSize Number of core and memory to be used for driver allocated in the specified Spark pool for the job, which will be used for overriding 'driverCores' and 'driverMemory' of the spark job definition you provide. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
executorSize Number of core and memory to be used for executors allocated in the specified Spark pool for the job, which will be used for overriding 'executorCores' and 'executorMemory' of the spark job definition you provide. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
file The main file used for the job, which will override the 'file' of the spark job definition you provide. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
files (Deprecated. Please use pythonCodeReference and filesV2) Additional files used for reference in the main definition file, which will override the 'files' of the spark job definition you provide. any[]
filesV2 Additional files used for reference in the main definition file, which will override the 'jars' and 'files' of the spark job definition you provide. any[]
numExecutors Number of executors to launch for this job, which will override the 'numExecutors' of the spark job definition you provide. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
pythonCodeReference Additional python code files used for reference in the main definition file, which will override the 'pyFiles' of the spark job definition you provide. any[]
scanFolder Scanning subfolders from the root folder of the main definition file, these files will be added as reference files. The folders named 'jars', 'pyFiles', 'files' or 'archives' will be scanned, and the folders name are case sensitive. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
sparkConfig Spark configuration property. SynapseSparkJobActivityTypePropertiesSparkConfig
sparkJob Synapse spark job reference. SynapseSparkJobReference (required)
targetBigDataPool The name of the big data pool which will be used to execute the spark batch job, which will override the 'targetBigDataPool' of the spark job definition you provide. BigDataPoolParametrizationReference
targetSparkConfiguration The spark configuration of the spark job. SparkConfigurationParametrizationReference

SynapseSparkJobActivityTypePropertiesSparkConfig

Name Description Value
{customized property} For Bicep, you can use the any() function.

SynapseSparkJobReference

Name Description Value
referenceName Reference spark job name. Expression with resultType string. For Bicep, you can use the any() function.(required)
type Synapse spark job reference type. 'SparkJobDefinitionReference' (required)

BigDataPoolParametrizationReference

Name Description Value
referenceName Reference big data pool name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
type Big data pool reference type. 'BigDataPoolReference' (required)

SparkConfigurationParametrizationReference

Name Description Value
referenceName Reference spark configuration name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
type Spark configuration reference type. 'SparkConfigurationReference' (required)

SqlServerStoredProcedureActivity

Name Description Value
type Type of activity. 'SqlServerStoredProcedure' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties SQL stored procedure activity properties. SqlServerStoredProcedureActivityTypeProperties (required)

SqlServerStoredProcedureActivityTypeProperties

Name Description Value
storedProcedureName Stored procedure name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". For Bicep, you can use the any() function.

SwitchActivity

Name Description Value
type Type of activity. 'Switch' (required)
typeProperties Switch activity properties. SwitchActivityTypeProperties (required)

SwitchActivityTypeProperties

Name Description Value
cases List of cases that correspond to expected values of the 'on' property. This is an optional property and if not provided, the activity will execute activities provided in defaultActivities. SwitchCase[]
defaultActivities List of activities to execute if no case condition is satisfied. This is an optional property and if not provided, the activity will exit without any action. Activity[]
on An expression that would evaluate to a string or integer. This is used to determine the block of activities in cases that will be executed. Expression (required)

SwitchCase

Name Description Value
activities List of activities to execute for satisfied case condition. Activity[]
value Expected value that satisfies the expression result of the 'on' property. string

SynapseNotebookActivity

Name Description Value
type Type of activity. 'SynapseNotebook' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Execute Synapse notebook activity properties. SynapseNotebookActivityTypeProperties (required)

SynapseNotebookActivityTypeProperties

Name Description Value
conf Spark configuration properties, which will override the 'conf' of the notebook you provide. For Bicep, you can use the any() function.
configurationType The type of the spark config. 'Artifact'
'Customized'
'Default'
driverSize Number of core and memory to be used for driver allocated in the specified Spark pool for the session, which will be used for overriding 'driverCores' and 'driverMemory' of the notebook you provide. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
executorSize Number of core and memory to be used for executors allocated in the specified Spark pool for the session, which will be used for overriding 'executorCores' and 'executorMemory' of the notebook you provide. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
notebook Synapse notebook reference. SynapseNotebookReference (required)
numExecutors Number of executors to launch for this session, which will override the 'numExecutors' of the notebook you provide. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
parameters Notebook parameters. SynapseNotebookActivityTypePropertiesParameters
sparkConfig Spark configuration property. SynapseNotebookActivityTypePropertiesSparkConfig
sparkPool The name of the big data pool which will be used to execute the notebook. BigDataPoolParametrizationReference
targetSparkConfiguration The spark configuration of the spark job. SparkConfigurationParametrizationReference

SynapseNotebookReference

Name Description Value
referenceName Reference notebook name. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)
type Synapse notebook reference type. 'NotebookReference' (required)

SynapseNotebookActivityTypePropertiesParameters

Name Description Value
{customized property} NotebookParameter

NotebookParameter

Name Description Value
type Notebook parameter type. 'bool'
'float'
'int'
'string'
value Notebook parameter value. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

SynapseNotebookActivityTypePropertiesSparkConfig

Name Description Value
{customized property} For Bicep, you can use the any() function.

UntilActivity

Name Description Value
type Type of activity. 'Until' (required)
typeProperties Until activity properties. UntilActivityTypeProperties (required)

UntilActivityTypeProperties

Name Description Value
activities List of activities to execute. Activity[] (required)
expression An expression that would evaluate to Boolean. The loop will continue until this expression evaluates to true Expression (required)
timeout Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

ValidationActivity

Name Description Value
type Type of activity. 'Validation' (required)
typeProperties Validation activity properties. ValidationActivityTypeProperties (required)

ValidationActivityTypeProperties

Name Description Value
childItems Can be used if dataset points to a folder. If set to true, the folder must have at least one file. If set to false, the folder must be empty. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
dataset Validation activity dataset reference. DatasetReference (required)
minimumSize Can be used if dataset points to a file. The file must be greater than or equal in size to the value specified. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
sleep A delay in seconds between validation attempts. If no value is specified, 10 seconds will be used as the default. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.
timeout Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). For Bicep, you can use the any() function.

WaitActivity

Name Description Value
type Type of activity. 'Wait' (required)
typeProperties Wait activity properties. WaitActivityTypeProperties (required)

WaitActivityTypeProperties

Name Description Value
waitTimeInSeconds Duration in seconds. Type: integer (or Expression with resultType integer). For Bicep, you can use the any() function.(required)

WebActivity

Name Description Value
type Type of activity. 'WebActivity' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Web activity properties. WebActivityTypeProperties (required)

WebActivityTypeProperties

Name Description Value
authentication Authentication method used for calling the endpoint. WebActivityAuthentication
body Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
connectVia The integration runtime reference. IntegrationRuntimeReference
datasets List of datasets passed to web endpoint. DatasetReference[]
disableCertValidation When set to true, Certificate validation will be disabled. bool
headers Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). WebActivityTypePropertiesHeaders
httpRequestTimeout Timeout for the HTTP request to get a response. Format is in TimeSpan (hh:mm:ss). This value is the timeout to get a response, not the activity timeout. The default value is 00:01:00 (1 minute). The range is from 1 to 10 minutes For Bicep, you can use the any() function.
linkedServices List of linked services passed to web endpoint. LinkedServiceReference[]
method Rest API method for target endpoint. 'DELETE'
'GET'
'POST'
'PUT' (required)
turnOffAsync Option to disable invoking HTTP GET on location given in response header of a HTTP 202 Response. If set true, it stops invoking HTTP GET on http location given in response header. If set false then continues to invoke HTTP GET call on location given in http response headers. bool
url Web activity target endpoint and path. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)

WebActivityAuthentication

Name Description Value
credential The credential reference containing authentication information. CredentialReference
password Password for the PFX file or basic authentication / Secret when used for ServicePrincipal SecretBase
pfx Base64-encoded contents of a PFX file or Certificate when used for ServicePrincipal SecretBase
resource Resource for which Azure Auth token will be requested when using MSI Authentication. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
type Web activity authentication (Basic/ClientCertificate/MSI/ServicePrincipal) string
username Web activity authentication user name for basic authentication or ClientID when used for ServicePrincipal. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
userTenant TenantId for which Azure Auth token will be requested when using ServicePrincipal Authentication. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.

CredentialReference

Name Description Value
referenceName Reference credential name. string (required)
type Credential reference type. 'CredentialReference' (required)
{customized property} For Bicep, you can use the any() function.

WebActivityTypePropertiesHeaders

Name Description Value
{customized property} string

WebHookActivity

Name Description Value
type Type of activity. 'WebHook' (required)
policy Activity policy. SecureInputOutputPolicy
typeProperties WebHook activity properties. WebHookActivityTypeProperties (required)

WebHookActivityTypeProperties

Name Description Value
authentication Authentication method used for calling the endpoint. WebActivityAuthentication
body Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). For Bicep, you can use the any() function.
headers Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). WebHookActivityTypePropertiesHeaders
method Rest API method for target endpoint. 'POST' (required)
reportStatusOnCallBack When set to true, statusCode, output and error in callback request body will be consumed by activity. The activity can be marked as failed by setting statusCode >= 400 in callback request. Default is false. Type: boolean (or Expression with resultType boolean). For Bicep, you can use the any() function.
timeout The timeout within which the webhook should be called back. If there is no value specified, it defaults to 10 minutes. Type: string. Pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). string
url WebHook activity target endpoint and path. Type: string (or Expression with resultType string). For Bicep, you can use the any() function.(required)

WebHookActivityTypePropertiesHeaders

Name Description Value
{customized property} string

PipelineFolder

Name Description Value
name The name of the folder that this Pipeline is in. string

ParameterDefinitionSpecification

Name Description Value
{customized property} ParameterSpecification

ParameterSpecification

Name Description Value
defaultValue Default value of parameter. For Bicep, you can use the any() function.
type Parameter type. 'Array'
'Bool'
'Float'
'Int'
'Object'
'SecureString'
'String' (required)

PipelinePolicy

Name Description Value
elapsedTimeMetric Pipeline ElapsedTime Metric Policy. PipelineElapsedTimeMetricPolicy

PipelineElapsedTimeMetricPolicy

Name Description Value
duration TimeSpan value, after which an Azure Monitoring Metric is fired. For Bicep, you can use the any() function.

PipelineRunDimensions

Name Description Value
{customized property} For Bicep, you can use the any() function.

VariableDefinitionSpecification

Name Description Value
{customized property} VariableSpecification

VariableSpecification

Name Description Value
defaultValue Default value of variable. For Bicep, you can use the any() function.
type Variable type. 'Array'
'Bool'
'String' (required)

Quickstart templates

The following quickstart templates deploy this resource type.

Template Description
More is possible with Azure Data Factory - One click to try Azure Data Factory

Deploy to Azure
This template creates a data factory pipeline for a copy activity from Azure Blob into another Azure Blob
Create a Data Factory, copy from Salesforce to Blob

Deploy to Azure
This template uses Azure Data Factory to perform data movement from Salesforce to Azure Blobs
Bulk copy using Azure Data Factory

Deploy to Azure
This template creates a V2 data factory that copies data in bulk an Azure SQL database to an Azure SQL data warehouse.
Create a V2 data factory

Deploy to Azure
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to another folder in the storage.
Create a V2 data factory (MySQL)

Deploy to Azure
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to a table in an Azure Database for MySQL.
Create a V2 data factory (PostGre)

Deploy to Azure
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to a table in an Azure Database for PostgreSQL.
Create a V2 data factory (SQL)

Deploy to Azure
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to a table in an Azure SQL database.
Create a V2 data factory (SQL On-prem)

Deploy to Azure
This template creates a V2 data factory that copies data from an on-premises SQL Server to an Azure blob storage
Create a V2 data factory (Spark)

Deploy to Azure
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to another folder in the storage.

ARM template resource definition

The factories/pipelines resource type can be deployed with operations that target:

For a list of changed properties in each API version, see change log.

Resource format

To create a Microsoft.DataFactory/factories/pipelines resource, add the following JSON to your template.

{
  "type": "Microsoft.DataFactory/factories/pipelines",
  "apiVersion": "2018-06-01",
  "name": "string",
  "properties": {
    "activities": [
      {
        "dependsOn": [
          {
            "activity": "string",
            "dependencyConditions": [ "string" ],
            "{customized property}": {}
          }
        ],
        "description": "string",
        "name": "string",
        "onInactiveMarkAs": "string",
        "state": "string",
        "userProperties": [
          {
            "name": "string",
            "value": {}
          }
        ],
        "type": "string"
        // For remaining properties, see Activity objects
      }
    ],
    "annotations": [ object ],
    "concurrency": "int",
    "description": "string",
    "folder": {
      "name": "string"
    },
    "parameters": {
      "{customized property}": {
        "defaultValue": {},
        "type": "string"
      }
    },
    "policy": {
      "elapsedTimeMetric": {
        "duration": {}
      }
    },
    "runDimensions": {
      "{customized property}": {}
    },
    "variables": {
      "{customized property}": {
        "defaultValue": {},
        "type": "string"
      }
    }
  }
}

Activity objects

Set the type property to specify the type of object.

For AppendVariable, use:

  "type": "AppendVariable",
  "typeProperties": {
    "value": {},
    "variableName": "string"
  }

For AzureDataExplorerCommand, use:

  "type": "AzureDataExplorerCommand",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "command": {},
    "commandTimeout": {}
  }

For AzureFunctionActivity, use:

  "type": "AzureFunctionActivity",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "body": {},
    "functionName": {},
    "headers": {
      "{customized property}": "string"
    },
    "method": "string"
  }

For AzureMLBatchExecution, use:

  "type": "AzureMLBatchExecution",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "globalParameters": {
      "{customized property}": {}
    },
    "webServiceInputs": {
      "{customized property}": {
        "filePath": {},
        "linkedServiceName": {
          "parameters": {
            "{customized property}": {}
          },
          "referenceName": "string",
          "type": "LinkedServiceReference"
        }
      }
    },
    "webServiceOutputs": {
      "{customized property}": {
        "filePath": {},
        "linkedServiceName": {
          "parameters": {
            "{customized property}": {}
          },
          "referenceName": "string",
          "type": "LinkedServiceReference"
        }
      }
    }
  }

For AzureMLExecutePipeline, use:

  "type": "AzureMLExecutePipeline",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "continueOnStepFailure": {},
    "dataPathAssignments": {},
    "experimentName": {},
    "mlParentRunId": {},
    "mlPipelineEndpointId": {},
    "mlPipelineId": {},
    "mlPipelineParameters": {},
    "version": {}
  }

For AzureMLUpdateResource, use:

  "type": "AzureMLUpdateResource",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "trainedModelFilePath": {},
    "trainedModelLinkedServiceName": {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "LinkedServiceReference"
    },
    "trainedModelName": {}
  }

For Copy, use:

  "type": "Copy",
  "inputs": [
    {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "DatasetReference"
    }
  ],
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "outputs": [
    {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "DatasetReference"
    }
  ],
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "dataIntegrationUnits": {},
    "enableSkipIncompatibleRow": {},
    "enableStaging": {},
    "logSettings": {
      "copyActivityLogSettings": {
        "enableReliableLogging": {},
        "logLevel": {}
      },
      "enableCopyActivityLog": {},
      "logLocationSettings": {
        "linkedServiceName": {
          "parameters": {
            "{customized property}": {}
          },
          "referenceName": "string",
          "type": "LinkedServiceReference"
        },
        "path": {}
      }
    },
    "logStorageSettings": {
      "enableReliableLogging": {},
      "linkedServiceName": {
        "parameters": {
          "{customized property}": {}
        },
        "referenceName": "string",
        "type": "LinkedServiceReference"
      },
      "logLevel": {},
      "path": {},
      "{customized property}": {}
    },
    "parallelCopies": {},
    "preserve": [ object ],
    "preserveRules": [ object ],
    "redirectIncompatibleRowSettings": {
      "linkedServiceName": {},
      "path": {},
      "{customized property}": {}
    },
    "sink": {
      "disableMetricsCollection": {},
      "maxConcurrentConnections": {},
      "sinkRetryCount": {},
      "sinkRetryWait": {},
      "writeBatchSize": {},
      "writeBatchTimeout": {},
      "type": "string"
      // For remaining properties, see CopySink objects
    },
    "skipErrorFile": {
      "dataInconsistency": {},
      "fileMissing": {}
    },
    "source": {
      "disableMetricsCollection": {},
      "maxConcurrentConnections": {},
      "sourceRetryCount": {},
      "sourceRetryWait": {},
      "type": "string"
      // For remaining properties, see CopySource objects
    },
    "stagingSettings": {
      "enableCompression": {},
      "linkedServiceName": {
        "parameters": {
          "{customized property}": {}
        },
        "referenceName": "string",
        "type": "LinkedServiceReference"
      },
      "path": {},
      "{customized property}": {}
    },
    "translator": {},
    "validateDataConsistency": {}
  }

For Custom, use:

  "type": "Custom",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "autoUserSpecification": {},
    "command": {},
    "extendedProperties": {
      "{customized property}": {}
    },
    "folderPath": {},
    "referenceObjects": {
      "datasets": [
        {
          "parameters": {
            "{customized property}": {}
          },
          "referenceName": "string",
          "type": "DatasetReference"
        }
      ],
      "linkedServices": [
        {
          "parameters": {
            "{customized property}": {}
          },
          "referenceName": "string",
          "type": "LinkedServiceReference"
        }
      ]
    },
    "resourceLinkedService": {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "LinkedServiceReference"
    },
    "retentionTimeInDays": {}
  }

For DatabricksNotebook, use:

  "type": "DatabricksNotebook",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "baseParameters": {
      "{customized property}": {}
    },
    "libraries": [
      {
        "{customized property}": {}
      }
    ],
    "notebookPath": {}
  }

For DatabricksSparkJar, use:

  "type": "DatabricksSparkJar",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "libraries": [
      {
        "{customized property}": {}
      }
    ],
    "mainClassName": {},
    "parameters": [ object ]
  }

For DatabricksSparkPython, use:

  "type": "DatabricksSparkPython",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "libraries": [
      {
        "{customized property}": {}
      }
    ],
    "parameters": [ object ],
    "pythonFile": {}
  }

For DataLakeAnalyticsU-SQL, use:

  "type": "DataLakeAnalyticsU-SQL",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "compilationMode": {},
    "degreeOfParallelism": {},
    "parameters": {
      "{customized property}": {}
    },
    "priority": {},
    "runtimeVersion": {},
    "scriptLinkedService": {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "LinkedServiceReference"
    },
    "scriptPath": {}
  }

For Delete, use:

  "type": "Delete",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "dataset": {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "DatasetReference"
    },
    "enableLogging": {},
    "logStorageSettings": {
      "enableReliableLogging": {},
      "linkedServiceName": {
        "parameters": {
          "{customized property}": {}
        },
        "referenceName": "string",
        "type": "LinkedServiceReference"
      },
      "logLevel": {},
      "path": {},
      "{customized property}": {}
    },
    "maxConcurrentConnections": "int",
    "recursive": {},
    "storeSettings": {
      "disableMetricsCollection": {},
      "maxConcurrentConnections": {},
      "type": "string"
      // For remaining properties, see StoreReadSettings objects
    }
  }

For ExecuteDataFlow, use:

  "type": "ExecuteDataFlow",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "compute": {
      "computeType": {},
      "coreCount": {}
    },
    "continueOnError": {},
    "dataFlow": {
      "datasetParameters": {},
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "DataFlowReference",
      "{customized property}": {}
    },
    "integrationRuntime": {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "IntegrationRuntimeReference"
    },
    "runConcurrently": {},
    "sourceStagingConcurrency": {},
    "staging": {
      "folderPath": {},
      "linkedService": {
        "parameters": {
          "{customized property}": {}
        },
        "referenceName": "string",
        "type": "LinkedServiceReference"
      }
    },
    "traceLevel": {}
  }

For ExecutePipeline, use:

  "type": "ExecutePipeline",
  "policy": {
    "secureInput": "bool",
    "{customized property}": {}
  },
  "typeProperties": {
    "parameters": {
      "{customized property}": {}
    },
    "pipeline": {
      "name": "string",
      "referenceName": "string",
      "type": "PipelineReference"
    },
    "waitOnCompletion": "bool"
  }

For ExecuteSSISPackage, use:

  "type": "ExecuteSSISPackage",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "connectVia": {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "IntegrationRuntimeReference"
    },
    "environmentPath": {},
    "executionCredential": {
      "domain": {},
      "password": {
        "type": "string",
        "value": "string"
      },
      "userName": {}
    },
    "loggingLevel": {},
    "logLocation": {
      "logPath": {},
      "type": "File",
      "typeProperties": {
        "accessCredential": {
          "domain": {},
          "password": {
            "type": "string"
            // For remaining properties, see SecretBase objects
          },
          "userName": {}
        },
        "logRefreshInterval": {}
      }
    },
    "packageConnectionManagers": {
      "{customized property}": {
        "{customized property}": {
          "value": {}
        }
      }
    },
    "packageLocation": {
      "packagePath": {},
      "type": "string",
      "typeProperties": {
        "accessCredential": {
          "domain": {},
          "password": {
            "type": "string"
            // For remaining properties, see SecretBase objects
          },
          "userName": {}
        },
        "childPackages": [
          {
            "packageContent": {},
            "packageLastModifiedDate": "string",
            "packageName": "string",
            "packagePath": {}
          }
        ],
        "configurationAccessCredential": {
          "domain": {},
          "password": {
            "type": "string"
            // For remaining properties, see SecretBase objects
          },
          "userName": {}
        },
        "configurationPath": {},
        "packageContent": {},
        "packageLastModifiedDate": "string",
        "packageName": "string",
        "packagePassword": {
          "type": "string"
          // For remaining properties, see SecretBase objects
        }
      }
    },
    "packageParameters": {
      "{customized property}": {
        "value": {}
      }
    },
    "projectConnectionManagers": {
      "{customized property}": {
        "{customized property}": {
          "value": {}
        }
      }
    },
    "projectParameters": {
      "{customized property}": {
        "value": {}
      }
    },
    "propertyOverrides": {
      "{customized property}": {
        "isSensitive": "bool",
        "value": {}
      }
    },
    "runtime": {}
  }

For ExecuteWranglingDataflow, use:

  "type": "ExecuteWranglingDataflow",
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "compute": {
      "computeType": {},
      "coreCount": {}
    },
    "continueOnError": {},
    "dataFlow": {
      "datasetParameters": {},
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "DataFlowReference",
      "{customized property}": {}
    },
    "integrationRuntime": {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "IntegrationRuntimeReference"
    },
    "queries": [
      {
        "dataflowSinks": [
          {
            "dataset": {
              "parameters": {
                "{customized property}": {}
              },
              "referenceName": "string",
              "type": "DatasetReference"
            },
            "description": "string",
            "flowlet": {
              "datasetParameters": {},
              "parameters": {
                "{customized property}": {}
              },
              "referenceName": "string",
              "type": "DataFlowReference",
              "{customized property}": {}
            },
            "linkedService": {
              "parameters": {
                "{customized property}": {}
              },
              "referenceName": "string",
              "type": "LinkedServiceReference"
            },
            "name": "string",
            "rejectedDataLinkedService": {
              "parameters": {
                "{customized property}": {}
              },
              "referenceName": "string",
              "type": "LinkedServiceReference"
            },
            "schemaLinkedService": {
              "parameters": {
                "{customized property}": {}
              },
              "referenceName": "string",
              "type": "LinkedServiceReference"
            },
            "script": "string"
          }
        ],
        "queryName": "string"
      }
    ],
    "runConcurrently": {},
    "sinks": {
      "{customized property}": {
        "dataset": {
          "parameters": {
            "{customized property}": {}
          },
          "referenceName": "string",
          "type": "DatasetReference"
        },
        "description": "string",
        "flowlet": {
          "datasetParameters": {},
          "parameters": {
            "{customized property}": {}
          },
          "referenceName": "string",
          "type": "DataFlowReference",
          "{customized property}": {}
        },
        "linkedService": {
          "parameters": {
            "{customized property}": {}
          },
          "referenceName": "string",
          "type": "LinkedServiceReference"
        },
        "name": "string",
        "rejectedDataLinkedService": {
          "parameters": {
            "{customized property}": {}
          },
          "referenceName": "string",
          "type": "LinkedServiceReference"
        },
        "schemaLinkedService": {
          "parameters": {
            "{customized property}": {}
          },
          "referenceName": "string",
          "type": "LinkedServiceReference"
        },
        "script": "string"
      }
    },
    "sourceStagingConcurrency": {},
    "staging": {
      "folderPath": {},
      "linkedService": {
        "parameters": {
          "{customized property}": {}
        },
        "referenceName": "string",
        "type": "LinkedServiceReference"
      }
    },
    "traceLevel": {}
  }

For Fail, use:

  "type": "Fail",
  "typeProperties": {
    "errorCode": {},
    "message": {}
  }

For Filter, use:

  "type": "Filter",
  "typeProperties": {
    "condition": {
      "type": "Expression",
      "value": "string"
    },
    "items": {
      "type": "Expression",
      "value": "string"
    }
  }

For ForEach, use:

  "type": "ForEach",
  "typeProperties": {
    "activities": [
      {
        "dependsOn": [
          {
            "activity": "string",
            "dependencyConditions": [ "string" ],
            "{customized property}": {}
          }
        ],
        "description": "string",
        "name": "string",
        "onInactiveMarkAs": "string",
        "state": "string",
        "userProperties": [
          {
            "name": "string",
            "value": {}
          }
        ],
        "type": "string"
        // For remaining properties, see Activity objects
      }
    ],
    "batchCount": "int",
    "isSequential": "bool",
    "items": {
      "type": "Expression",
      "value": "string"
    }
  }

For GetMetadata, use:

  "type": "GetMetadata",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "dataset": {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "DatasetReference"
    },
    "fieldList": [ object ],
    "formatSettings": {
      "type": "string"
      // For remaining properties, see FormatReadSettings objects
    },
    "storeSettings": {
      "disableMetricsCollection": {},
      "maxConcurrentConnections": {},
      "type": "string"
      // For remaining properties, see StoreReadSettings objects
    }
  }

For HDInsightHive, use:

  "type": "HDInsightHive",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "arguments": [ object ],
    "defines": {
      "{customized property}": {}
    },
    "getDebugInfo": "string",
    "queryTimeout": "int",
    "scriptLinkedService": {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "LinkedServiceReference"
    },
    "scriptPath": {},
    "storageLinkedServices": [
      {
        "parameters": {
          "{customized property}": {}
        },
        "referenceName": "string",
        "type": "LinkedServiceReference"
      }
    ],
    "variables": {
      "{customized property}": {}
    }
  }

For HDInsightMapReduce, use:

  "type": "HDInsightMapReduce",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "arguments": [ object ],
    "className": {},
    "defines": {
      "{customized property}": {}
    },
    "getDebugInfo": "string",
    "jarFilePath": {},
    "jarLibs": [ object ],
    "jarLinkedService": {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "LinkedServiceReference"
    },
    "storageLinkedServices": [
      {
        "parameters": {
          "{customized property}": {}
        },
        "referenceName": "string",
        "type": "LinkedServiceReference"
      }
    ]
  }

For HDInsightPig, use:

  "type": "HDInsightPig",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "arguments": {},
    "defines": {
      "{customized property}": {}
    },
    "getDebugInfo": "string",
    "scriptLinkedService": {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "LinkedServiceReference"
    },
    "scriptPath": {},
    "storageLinkedServices": [
      {
        "parameters": {
          "{customized property}": {}
        },
        "referenceName": "string",
        "type": "LinkedServiceReference"
      }
    ]
  }

For HDInsightSpark, use:

  "type": "HDInsightSpark",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "arguments": [ object ],
    "className": "string",
    "entryFilePath": {},
    "getDebugInfo": "string",
    "proxyUser": {},
    "rootPath": {},
    "sparkConfig": {
      "{customized property}": {}
    },
    "sparkJobLinkedService": {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "LinkedServiceReference"
    }
  }

For HDInsightStreaming, use:

  "type": "HDInsightStreaming",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "arguments": [ object ],
    "combiner": {},
    "commandEnvironment": [ object ],
    "defines": {
      "{customized property}": {}
    },
    "fileLinkedService": {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "LinkedServiceReference"
    },
    "filePaths": [ object ],
    "getDebugInfo": "string",
    "input": {},
    "mapper": {},
    "output": {},
    "reducer": {},
    "storageLinkedServices": [
      {
        "parameters": {
          "{customized property}": {}
        },
        "referenceName": "string",
        "type": "LinkedServiceReference"
      }
    ]
  }

For IfCondition, use:

  "type": "IfCondition",
  "typeProperties": {
    "expression": {
      "type": "Expression",
      "value": "string"
    },
    "ifFalseActivities": [
      {
        "dependsOn": [
          {
            "activity": "string",
            "dependencyConditions": [ "string" ],
            "{customized property}": {}
          }
        ],
        "description": "string",
        "name": "string",
        "onInactiveMarkAs": "string",
        "state": "string",
        "userProperties": [
          {
            "name": "string",
            "value": {}
          }
        ],
        "type": "string"
        // For remaining properties, see Activity objects
      }
    ],
    "ifTrueActivities": [
      {
        "dependsOn": [
          {
            "activity": "string",
            "dependencyConditions": [ "string" ],
            "{customized property}": {}
          }
        ],
        "description": "string",
        "name": "string",
        "onInactiveMarkAs": "string",
        "state": "string",
        "userProperties": [
          {
            "name": "string",
            "value": {}
          }
        ],
        "type": "string"
        // For remaining properties, see Activity objects
      }
    ]
  }

For Lookup, use:

  "type": "Lookup",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "dataset": {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "DatasetReference"
    },
    "firstRowOnly": {},
    "source": {
      "disableMetricsCollection": {},
      "maxConcurrentConnections": {},
      "sourceRetryCount": {},
      "sourceRetryWait": {},
      "type": "string"
      // For remaining properties, see CopySource objects
    }
  }

For Script, use:

  "type": "Script",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "logSettings": {
      "logDestination": "string",
      "logLocationSettings": {
        "linkedServiceName": {
          "parameters": {
            "{customized property}": {}
          },
          "referenceName": "string",
          "type": "LinkedServiceReference"
        },
        "path": {}
      }
    },
    "scriptBlockExecutionTimeout": {},
    "scripts": [
      {
        "parameters": [
          {
            "direction": "string",
            "name": {},
            "size": "int",
            "type": "string",
            "value": {}
          }
        ],
        "text": {},
        "type": "string"
      }
    ]
  }

For SetVariable, use:

  "type": "SetVariable",
  "policy": {
    "secureInput": "bool",
    "secureOutput": "bool"
  },
  "typeProperties": {
    "setSystemVariable": "bool",
    "value": {},
    "variableName": "string"
  }

For SparkJob, use:

  "type": "SparkJob",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "args": [ object ],
    "className": {},
    "conf": {},
    "configurationType": "string",
    "driverSize": {},
    "executorSize": {},
    "file": {},
    "files": [ object ],
    "filesV2": [ object ],
    "numExecutors": {},
    "pythonCodeReference": [ object ],
    "scanFolder": {},
    "sparkConfig": {
      "{customized property}": {}
    },
    "sparkJob": {
      "referenceName": {},
      "type": "SparkJobDefinitionReference"
    },
    "targetBigDataPool": {
      "referenceName": {},
      "type": "BigDataPoolReference"
    },
    "targetSparkConfiguration": {
      "referenceName": {},
      "type": "SparkConfigurationReference"
    }
  }

For SqlServerStoredProcedure, use:

  "type": "SqlServerStoredProcedure",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "storedProcedureName": {},
    "storedProcedureParameters": {}
  }

For Switch, use:

  "type": "Switch",
  "typeProperties": {
    "cases": [
      {
        "activities": [
          {
            "dependsOn": [
              {
                "activity": "string",
                "dependencyConditions": [ "string" ],
                "{customized property}": {}
              }
            ],
            "description": "string",
            "name": "string",
            "onInactiveMarkAs": "string",
            "state": "string",
            "userProperties": [
              {
                "name": "string",
                "value": {}
              }
            ],
            "type": "string"
            // For remaining properties, see Activity objects
          }
        ],
        "value": "string"
      }
    ],
    "defaultActivities": [
      {
        "dependsOn": [
          {
            "activity": "string",
            "dependencyConditions": [ "string" ],
            "{customized property}": {}
          }
        ],
        "description": "string",
        "name": "string",
        "onInactiveMarkAs": "string",
        "state": "string",
        "userProperties": [
          {
            "name": "string",
            "value": {}
          }
        ],
        "type": "string"
        // For remaining properties, see Activity objects
      }
    ],
    "on": {
      "type": "Expression",
      "value": "string"
    }
  }

For SynapseNotebook, use:

  "type": "SynapseNotebook",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "conf": {},
    "configurationType": "string",
    "driverSize": {},
    "executorSize": {},
    "notebook": {
      "referenceName": {},
      "type": "NotebookReference"
    },
    "numExecutors": {},
    "parameters": {
      "{customized property}": {
        "type": "string",
        "value": {}
      }
    },
    "sparkConfig": {
      "{customized property}": {}
    },
    "sparkPool": {
      "referenceName": {},
      "type": "BigDataPoolReference"
    },
    "targetSparkConfiguration": {
      "referenceName": {},
      "type": "SparkConfigurationReference"
    }
  }

For Until, use:

  "type": "Until",
  "typeProperties": {
    "activities": [
      {
        "dependsOn": [
          {
            "activity": "string",
            "dependencyConditions": [ "string" ],
            "{customized property}": {}
          }
        ],
        "description": "string",
        "name": "string",
        "onInactiveMarkAs": "string",
        "state": "string",
        "userProperties": [
          {
            "name": "string",
            "value": {}
          }
        ],
        "type": "string"
        // For remaining properties, see Activity objects
      }
    ],
    "expression": {
      "type": "Expression",
      "value": "string"
    },
    "timeout": {}
  }

For Validation, use:

  "type": "Validation",
  "typeProperties": {
    "childItems": {},
    "dataset": {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "DatasetReference"
    },
    "minimumSize": {},
    "sleep": {},
    "timeout": {}
  }

For Wait, use:

  "type": "Wait",
  "typeProperties": {
    "waitTimeInSeconds": {}
  }

For WebActivity, use:

  "type": "WebActivity",
  "linkedServiceName": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  },
  "policy": {
    "retry": {},
    "retryIntervalInSeconds": "int",
    "secureInput": "bool",
    "secureOutput": "bool",
    "timeout": {},
    "{customized property}": {}
  },
  "typeProperties": {
    "authentication": {
      "credential": {
        "referenceName": "string",
        "type": "CredentialReference",
        "{customized property}": {}
      },
      "password": {
        "type": "string"
        // For remaining properties, see SecretBase objects
      },
      "pfx": {
        "type": "string"
        // For remaining properties, see SecretBase objects
      },
      "resource": {},
      "type": "string",
      "username": {},
      "userTenant": {}
    },
    "body": {},
    "connectVia": {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "IntegrationRuntimeReference"
    },
    "datasets": [
      {
        "parameters": {
          "{customized property}": {}
        },
        "referenceName": "string",
        "type": "DatasetReference"
      }
    ],
    "disableCertValidation": "bool",
    "headers": {
      "{customized property}": "string"
    },
    "httpRequestTimeout": {},
    "linkedServices": [
      {
        "parameters": {
          "{customized property}": {}
        },
        "referenceName": "string",
        "type": "LinkedServiceReference"
      }
    ],
    "method": "string",
    "turnOffAsync": "bool",
    "url": {}
  }

For WebHook, use:

  "type": "WebHook",
  "policy": {
    "secureInput": "bool",
    "secureOutput": "bool"
  },
  "typeProperties": {
    "authentication": {
      "credential": {
        "referenceName": "string",
        "type": "CredentialReference",
        "{customized property}": {}
      },
      "password": {
        "type": "string"
        // For remaining properties, see SecretBase objects
      },
      "pfx": {
        "type": "string"
        // For remaining properties, see SecretBase objects
      },
      "resource": {},
      "type": "string",
      "username": {},
      "userTenant": {}
    },
    "body": {},
    "headers": {
      "{customized property}": "string"
    },
    "method": "POST",
    "reportStatusOnCallBack": {},
    "timeout": "string",
    "url": {}
  }

CopySink objects

Set the type property to specify the type of object.

For AvroSink, use:

  "type": "AvroSink",
  "formatSettings": {
    "fileNamePrefix": {},
    "maxRowsPerFile": {},
    "recordName": "string",
    "recordNamespace": "string",
    "type": "string",
    "{customized property}": {}
  },
  "storeSettings": {
    "copyBehavior": {},
    "disableMetricsCollection": {},
    "maxConcurrentConnections": {},
    "metadata": [
      {
        "name": {},
        "value": {}
      }
    ],
    "type": "string"
    // For remaining properties, see StoreWriteSettings objects
  }

For AzureBlobFSSink, use:

  "type": "AzureBlobFSSink",
  "copyBehavior": {},
  "metadata": [
    {
      "name": {},
      "value": {}
    }
  ]

For AzureDatabricksDeltaLakeSink, use:

  "type": "AzureDatabricksDeltaLakeSink",
  "importSettings": {
    "dateFormat": {},
    "timestampFormat": {},
    "type": "string",
    "{customized property}": {}
  },
  "preCopyScript": {}

For AzureDataExplorerSink, use:

  "type": "AzureDataExplorerSink",
  "flushImmediately": {},
  "ingestionMappingAsJson": {},
  "ingestionMappingName": {}

For AzureDataLakeStoreSink, use:

  "type": "AzureDataLakeStoreSink",
  "copyBehavior": {},
  "enableAdlsSingleFileParallel": {}

For AzureMySqlSink, use:

  "type": "AzureMySqlSink",
  "preCopyScript": {}

For AzurePostgreSqlSink, use:

  "type": "AzurePostgreSqlSink",
  "preCopyScript": {}

For AzureQueueSink, use:

  "type": "AzureQueueSink"

For AzureSearchIndexSink, use:

  "type": "AzureSearchIndexSink",
  "writeBehavior": "string"

For AzureSqlSink, use:

  "type": "AzureSqlSink",
  "preCopyScript": {},
  "sqlWriterStoredProcedureName": {},
  "sqlWriterTableType": {},
  "sqlWriterUseTableLock": {},
  "storedProcedureParameters": {},
  "storedProcedureTableTypeParameterName": {},
  "tableOption": {},
  "upsertSettings": {
    "interimSchemaName": {},
    "keys": {},
    "useTempDB": {}
  },
  "writeBehavior": {}

For AzureTableSink, use:

  "type": "AzureTableSink",
  "azureTableDefaultPartitionKeyValue": {},
  "azureTableInsertType": {},
  "azureTablePartitionKeyName": {},
  "azureTableRowKeyName": {}

For BinarySink, use:

  "type": "BinarySink",
  "storeSettings": {
    "copyBehavior": {},
    "disableMetricsCollection": {},
    "maxConcurrentConnections": {},
    "metadata": [
      {
        "name": {},
        "value": {}
      }
    ],
    "type": "string"
    // For remaining properties, see StoreWriteSettings objects
  }

For BlobSink, use:

  "type": "BlobSink",
  "blobWriterAddHeader": {},
  "blobWriterDateTimeFormat": {},
  "blobWriterOverwriteFiles": {},
  "copyBehavior": {},
  "metadata": [
    {
      "name": {},
      "value": {}
    }
  ]

For CommonDataServiceForAppsSink, use:

  "type": "CommonDataServiceForAppsSink",
  "alternateKeyName": {},
  "ignoreNullValues": {},
  "writeBehavior": "Upsert"

For CosmosDbMongoDbApiSink, use:

  "type": "CosmosDbMongoDbApiSink",
  "writeBehavior": {}

For CosmosDbSqlApiSink, use:

  "type": "CosmosDbSqlApiSink",
  "writeBehavior": {}

For DelimitedTextSink, use:

  "type": "DelimitedTextSink",
  "formatSettings": {
    "fileExtension": {},
    "fileNamePrefix": {},
    "maxRowsPerFile": {},
    "quoteAllText": {},
    "type": "string",
    "{customized property}": {}
  },
  "storeSettings": {
    "copyBehavior": {},
    "disableMetricsCollection": {},
    "maxConcurrentConnections": {},
    "metadata": [
      {
        "name": {},
        "value": {}
      }
    ],
    "type": "string"
    // For remaining properties, see StoreWriteSettings objects
  }

For DocumentDbCollectionSink, use:

  "type": "DocumentDbCollectionSink",
  "nestingSeparator": {},
  "writeBehavior": {}

For DynamicsCrmSink, use:

  "type": "DynamicsCrmSink",
  "alternateKeyName": {},
  "ignoreNullValues": {},
  "writeBehavior": "Upsert"

For DynamicsSink, use:

  "type": "DynamicsSink",
  "alternateKeyName": {},
  "ignoreNullValues": {},
  "writeBehavior": "Upsert"

For FileSystemSink, use:

  "type": "FileSystemSink",
  "copyBehavior": {}

For InformixSink, use:

  "type": "InformixSink",
  "preCopyScript": {}

For JsonSink, use:

  "type": "JsonSink",
  "formatSettings": {
    "filePattern": {},
    "type": "string",
    "{customized property}": {}
  },
  "storeSettings": {
    "copyBehavior": {},
    "disableMetricsCollection": {},
    "maxConcurrentConnections": {},
    "metadata": [
      {
        "name": {},
        "value": {}
      }
    ],
    "type": "string"
    // For remaining properties, see StoreWriteSettings objects
  }

For LakeHouseTableSink, use:

  "type": "LakeHouseTableSink",
  "partitionNameList": {},
  "partitionOption": {},
  "tableActionOption": {}

For MicrosoftAccessSink, use:

  "type": "MicrosoftAccessSink",
  "preCopyScript": {}

For MongoDbAtlasSink, use:

  "type": "MongoDbAtlasSink",
  "writeBehavior": {}

For MongoDbV2Sink, use:

  "type": "MongoDbV2Sink",
  "writeBehavior": {}

For OdbcSink, use:

  "type": "OdbcSink",
  "preCopyScript": {}

For OracleSink, use:

  "type": "OracleSink",
  "preCopyScript": {}

For OrcSink, use:

  "type": "OrcSink",
  "formatSettings": {
    "fileNamePrefix": {},
    "maxRowsPerFile": {},
    "type": "string",
    "{customized property}": {}
  },
  "storeSettings": {
    "copyBehavior": {},
    "disableMetricsCollection": {},
    "maxConcurrentConnections": {},
    "metadata": [
      {
        "name": {},
        "value": {}
      }
    ],
    "type": "string"
    // For remaining properties, see StoreWriteSettings objects
  }

For ParquetSink, use:

  "type": "ParquetSink",
  "formatSettings": {
    "fileNamePrefix": {},
    "maxRowsPerFile": {},
    "type": "string",
    "{customized property}": {}
  },
  "storeSettings": {
    "copyBehavior": {},
    "disableMetricsCollection": {},
    "maxConcurrentConnections": {},
    "metadata": [
      {
        "name": {},
        "value": {}
      }
    ],
    "type": "string"
    // For remaining properties, see StoreWriteSettings objects
  }

For RestSink, use:

  "type": "RestSink",
  "additionalHeaders": {},
  "httpCompressionType": {},
  "httpRequestTimeout": {},
  "requestInterval": {},
  "requestMethod": {}

For SalesforceServiceCloudSink, use:

  "type": "SalesforceServiceCloudSink",
  "externalIdFieldName": {},
  "ignoreNullValues": {},
  "writeBehavior": "string"

For SalesforceServiceCloudV2Sink, use:

  "type": "SalesforceServiceCloudV2Sink",
  "externalIdFieldName": {},
  "ignoreNullValues": {},
  "writeBehavior": "string"

For SalesforceSink, use:

  "type": "SalesforceSink",
  "externalIdFieldName": {},
  "ignoreNullValues": {},
  "writeBehavior": "string"

For SalesforceV2Sink, use:

  "type": "SalesforceV2Sink",
  "externalIdFieldName": {},
  "ignoreNullValues": {},
  "writeBehavior": "string"

For SapCloudForCustomerSink, use:

  "type": "SapCloudForCustomerSink",
  "httpRequestTimeout": {},
  "writeBehavior": "string"

For SnowflakeSink, use:

  "type": "SnowflakeSink",
  "importSettings": {
    "additionalCopyOptions": {
      "{customized property}": {}
    },
    "additionalFormatOptions": {
      "{customized property}": {}
    },
    "type": "string",
    "{customized property}": {}
  },
  "preCopyScript": {}

For SnowflakeV2Sink, use:

  "type": "SnowflakeV2Sink",
  "importSettings": {
    "additionalCopyOptions": {
      "{customized property}": {}
    },
    "additionalFormatOptions": {
      "{customized property}": {}
    },
    "type": "string",
    "{customized property}": {}
  },
  "preCopyScript": {}

For SqlDWSink, use:

  "type": "SqlDWSink",
  "allowCopyCommand": {},
  "allowPolyBase": {},
  "copyCommandSettings": {
    "additionalOptions": {
      "{customized property}": "string"
    },
    "defaultValues": [
      {
        "columnName": {},
        "defaultValue": {}
      }
    ]
  },
  "polyBaseSettings": {
    "rejectSampleValue": {},
    "rejectType": "string",
    "rejectValue": {},
    "useTypeDefault": {},
    "{customized property}": {}
  },
  "preCopyScript": {},
  "sqlWriterUseTableLock": {},
  "tableOption": {},
  "upsertSettings": {
    "interimSchemaName": {},
    "keys": {}
  },
  "writeBehavior": {}

For SqlMISink, use:

  "type": "SqlMISink",
  "preCopyScript": {},
  "sqlWriterStoredProcedureName": {},
  "sqlWriterTableType": {},
  "sqlWriterUseTableLock": {},
  "storedProcedureParameters": {},
  "storedProcedureTableTypeParameterName": {},
  "tableOption": {},
  "upsertSettings": {
    "interimSchemaName": {},
    "keys": {},
    "useTempDB": {}
  },
  "writeBehavior": {}

For SqlServerSink, use:

  "type": "SqlServerSink",
  "preCopyScript": {},
  "sqlWriterStoredProcedureName": {},
  "sqlWriterTableType": {},
  "sqlWriterUseTableLock": {},
  "storedProcedureParameters": {},
  "storedProcedureTableTypeParameterName": {},
  "tableOption": {},
  "upsertSettings": {
    "interimSchemaName": {},
    "keys": {},
    "useTempDB": {}
  },
  "writeBehavior": {}

For SqlSink, use:

  "type": "SqlSink",
  "preCopyScript": {},
  "sqlWriterStoredProcedureName": {},
  "sqlWriterTableType": {},
  "sqlWriterUseTableLock": {},
  "storedProcedureParameters": {},
  "storedProcedureTableTypeParameterName": {},
  "tableOption": {},
  "upsertSettings": {
    "interimSchemaName": {},
    "keys": {},
    "useTempDB": {}
  },
  "writeBehavior": {}

For WarehouseSink, use:

  "type": "WarehouseSink",
  "allowCopyCommand": {},
  "copyCommandSettings": {
    "additionalOptions": {
      "{customized property}": "string"
    },
    "defaultValues": [
      {
        "columnName": {},
        "defaultValue": {}
      }
    ]
  },
  "preCopyScript": {},
  "tableOption": {},
  "writeBehavior": {}

StoreWriteSettings objects

Set the type property to specify the type of object.

For AzureBlobFSWriteSettings, use:

  "type": "AzureBlobFSWriteSettings",
  "blockSizeInMB": {}

For AzureBlobStorageWriteSettings, use:

  "type": "AzureBlobStorageWriteSettings",
  "blockSizeInMB": {}

For AzureDataLakeStoreWriteSettings, use:

  "type": "AzureDataLakeStoreWriteSettings",
  "expiryDateTime": {}

For AzureFileStorageWriteSettings, use:

  "type": "AzureFileStorageWriteSettings"

For FileServerWriteSettings, use:

  "type": "FileServerWriteSettings"

For LakeHouseWriteSettings, use:

  "type": "LakeHouseWriteSettings"

For SftpWriteSettings, use:

  "type": "SftpWriteSettings",
  "operationTimeout": {},
  "useTempFileRename": {}

CopySource objects

Set the type property to specify the type of object.

For AmazonMWSSource, use:

  "type": "AmazonMWSSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For AmazonRdsForOracleSource, use:

  "type": "AmazonRdsForOracleSource",
  "additionalColumns": {},
  "oracleReaderQuery": {},
  "partitionOption": {},
  "partitionSettings": {
    "partitionColumnName": {},
    "partitionLowerBound": {},
    "partitionNames": {},
    "partitionUpperBound": {}
  },
  "queryTimeout": {}

For AmazonRdsForSqlServerSource, use:

  "type": "AmazonRdsForSqlServerSource",
  "additionalColumns": {},
  "isolationLevel": {},
  "partitionOption": {},
  "partitionSettings": {
    "partitionColumnName": {},
    "partitionLowerBound": {},
    "partitionUpperBound": {}
  },
  "produceAdditionalTypes": {},
  "queryTimeout": {},
  "sqlReaderQuery": {},
  "sqlReaderStoredProcedureName": {},
  "storedProcedureParameters": {}

For AmazonRedshiftSource, use:

  "type": "AmazonRedshiftSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {},
  "redshiftUnloadSettings": {
    "bucketName": {},
    "s3LinkedServiceName": {
      "parameters": {
        "{customized property}": {}
      },
      "referenceName": "string",
      "type": "LinkedServiceReference"
    }
  }

For AvroSource, use:

  "type": "AvroSource",
  "additionalColumns": {},
  "storeSettings": {
    "disableMetricsCollection": {},
    "maxConcurrentConnections": {},
    "type": "string"
    // For remaining properties, see StoreReadSettings objects
  }

For AzureBlobFSSource, use:

  "type": "AzureBlobFSSource",
  "recursive": {},
  "skipHeaderLineCount": {},
  "treatEmptyAsNull": {}

For AzureDatabricksDeltaLakeSource, use:

  "type": "AzureDatabricksDeltaLakeSource",
  "exportSettings": {
    "dateFormat": {},
    "timestampFormat": {},
    "type": "string",
    "{customized property}": {}
  },
  "query": {}

For AzureDataExplorerSource, use:

  "type": "AzureDataExplorerSource",
  "additionalColumns": {},
  "noTruncation": {},
  "query": {},
  "queryTimeout": {}

For AzureDataLakeStoreSource, use:

  "type": "AzureDataLakeStoreSource",
  "recursive": {}

For AzureMariaDBSource, use:

  "type": "AzureMariaDBSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For AzureMySqlSource, use:

  "type": "AzureMySqlSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For AzurePostgreSqlSource, use:

  "type": "AzurePostgreSqlSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For AzureSqlSource, use:

  "type": "AzureSqlSource",
  "additionalColumns": {},
  "isolationLevel": {},
  "partitionOption": {},
  "partitionSettings": {
    "partitionColumnName": {},
    "partitionLowerBound": {},
    "partitionUpperBound": {}
  },
  "produceAdditionalTypes": {},
  "queryTimeout": {},
  "sqlReaderQuery": {},
  "sqlReaderStoredProcedureName": {},
  "storedProcedureParameters": {}

For AzureTableSource, use:

  "type": "AzureTableSource",
  "additionalColumns": {},
  "azureTableSourceIgnoreTableNotFound": {},
  "azureTableSourceQuery": {},
  "queryTimeout": {}

For BinarySource, use:

  "type": "BinarySource",
  "formatSettings": {
    "compressionProperties": {
      "type": "string"
      // For remaining properties, see CompressionReadSettings objects
    },
    "type": "string",
    "{customized property}": {}
  },
  "storeSettings": {
    "disableMetricsCollection": {},
    "maxConcurrentConnections": {},
    "type": "string"
    // For remaining properties, see StoreReadSettings objects
  }

For BlobSource, use:

  "type": "BlobSource",
  "recursive": {},
  "skipHeaderLineCount": {},
  "treatEmptyAsNull": {}

For CassandraSource, use:

  "type": "CassandraSource",
  "additionalColumns": {},
  "consistencyLevel": "string",
  "query": {},
  "queryTimeout": {}

For CommonDataServiceForAppsSource, use:

  "type": "CommonDataServiceForAppsSource",
  "additionalColumns": {},
  "query": {}

For ConcurSource, use:

  "type": "ConcurSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For CosmosDbMongoDbApiSource, use:

  "type": "CosmosDbMongoDbApiSource",
  "additionalColumns": {},
  "batchSize": {},
  "cursorMethods": {
    "limit": {},
    "project": {},
    "skip": {},
    "sort": {},
    "{customized property}": {}
  },
  "filter": {},
  "queryTimeout": {}

For CosmosDbSqlApiSource, use:

  "type": "CosmosDbSqlApiSource",
  "additionalColumns": {},
  "detectDatetime": {},
  "pageSize": {},
  "preferredRegions": {},
  "query": {}

For CouchbaseSource, use:

  "type": "CouchbaseSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For Db2Source, use:

  "type": "Db2Source",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For DelimitedTextSource, use:

  "type": "DelimitedTextSource",
  "additionalColumns": {},
  "formatSettings": {
    "compressionProperties": {
      "type": "string"
      // For remaining properties, see CompressionReadSettings objects
    },
    "skipLineCount": {},
    "type": "string",
    "{customized property}": {}
  },
  "storeSettings": {
    "disableMetricsCollection": {},
    "maxConcurrentConnections": {},
    "type": "string"
    // For remaining properties, see StoreReadSettings objects
  }

For DocumentDbCollectionSource, use:

  "type": "DocumentDbCollectionSource",
  "additionalColumns": {},
  "nestingSeparator": {},
  "query": {},
  "queryTimeout": {}

For DrillSource, use:

  "type": "DrillSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For DynamicsAXSource, use:

  "type": "DynamicsAXSource",
  "additionalColumns": {},
  "httpRequestTimeout": {},
  "query": {},
  "queryTimeout": {}

For DynamicsCrmSource, use:

  "type": "DynamicsCrmSource",
  "additionalColumns": {},
  "query": {}

For DynamicsSource, use:

  "type": "DynamicsSource",
  "additionalColumns": {},
  "query": {}

For EloquaSource, use:

  "type": "EloquaSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For ExcelSource, use:

  "type": "ExcelSource",
  "additionalColumns": {},
  "storeSettings": {
    "disableMetricsCollection": {},
    "maxConcurrentConnections": {},
    "type": "string"
    // For remaining properties, see StoreReadSettings objects
  }

For FileSystemSource, use:

  "type": "FileSystemSource",
  "additionalColumns": {},
  "recursive": {}

For GoogleAdWordsSource, use:

  "type": "GoogleAdWordsSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For GoogleBigQuerySource, use:

  "type": "GoogleBigQuerySource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For GoogleBigQueryV2Source, use:

  "type": "GoogleBigQueryV2Source",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For GreenplumSource, use:

  "type": "GreenplumSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For HBaseSource, use:

  "type": "HBaseSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For HdfsSource, use:

  "type": "HdfsSource",
  "distcpSettings": {
    "distcpOptions": {},
    "resourceManagerEndpoint": {},
    "tempScriptPath": {}
  },
  "recursive": {}

For HiveSource, use:

  "type": "HiveSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For HttpSource, use:

  "type": "HttpSource",
  "httpRequestTimeout": {}

For HubspotSource, use:

  "type": "HubspotSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For ImpalaSource, use:

  "type": "ImpalaSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For InformixSource, use:

  "type": "InformixSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For JiraSource, use:

  "type": "JiraSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For JsonSource, use:

  "type": "JsonSource",
  "additionalColumns": {},
  "formatSettings": {
    "compressionProperties": {
      "type": "string"
      // For remaining properties, see CompressionReadSettings objects
    },
    "type": "string",
    "{customized property}": {}
  },
  "storeSettings": {
    "disableMetricsCollection": {},
    "maxConcurrentConnections": {},
    "type": "string"
    // For remaining properties, see StoreReadSettings objects
  }

For LakeHouseTableSource, use:

  "type": "LakeHouseTableSource",
  "additionalColumns": {},
  "timestampAsOf": {},
  "versionAsOf": {}

For MagentoSource, use:

  "type": "MagentoSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For MariaDBSource, use:

  "type": "MariaDBSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For MarketoSource, use:

  "type": "MarketoSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For MicrosoftAccessSource, use:

  "type": "MicrosoftAccessSource",
  "additionalColumns": {},
  "query": {}

For MongoDbAtlasSource, use:

  "type": "MongoDbAtlasSource",
  "additionalColumns": {},
  "batchSize": {},
  "cursorMethods": {
    "limit": {},
    "project": {},
    "skip": {},
    "sort": {},
    "{customized property}": {}
  },
  "filter": {},
  "queryTimeout": {}

For MongoDbSource, use:

  "type": "MongoDbSource",
  "additionalColumns": {},
  "query": {}

For MongoDbV2Source, use:

  "type": "MongoDbV2Source",
  "additionalColumns": {},
  "batchSize": {},
  "cursorMethods": {
    "limit": {},
    "project": {},
    "skip": {},
    "sort": {},
    "{customized property}": {}
  },
  "filter": {},
  "queryTimeout": {}

For MySqlSource, use:

  "type": "MySqlSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For NetezzaSource, use:

  "type": "NetezzaSource",
  "additionalColumns": {},
  "partitionOption": {},
  "partitionSettings": {
    "partitionColumnName": {},
    "partitionLowerBound": {},
    "partitionUpperBound": {}
  },
  "query": {},
  "queryTimeout": {}

For ODataSource, use:

  "type": "ODataSource",
  "additionalColumns": {},
  "httpRequestTimeout": {},
  "query": {}

For OdbcSource, use:

  "type": "OdbcSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For Office365Source, use:

  "type": "Office365Source",
  "allowedGroups": {},
  "dateFilterColumn": {},
  "endTime": {},
  "outputColumns": {},
  "startTime": {},
  "userScopeFilterUri": {}

For OracleServiceCloudSource, use:

  "type": "OracleServiceCloudSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For OracleSource, use:

  "type": "OracleSource",
  "additionalColumns": {},
  "oracleReaderQuery": {},
  "partitionOption": {},
  "partitionSettings": {
    "partitionColumnName": {},
    "partitionLowerBound": {},
    "partitionNames": {},
    "partitionUpperBound": {}
  },
  "queryTimeout": {}

For OrcSource, use:

  "type": "OrcSource",
  "additionalColumns": {},
  "storeSettings": {
    "disableMetricsCollection": {},
    "maxConcurrentConnections": {},
    "type": "string"
    // For remaining properties, see StoreReadSettings objects
  }

For ParquetSource, use:

  "type": "ParquetSource",
  "additionalColumns": {},
  "formatSettings": {
    "compressionProperties": {
      "type": "string"
      // For remaining properties, see CompressionReadSettings objects
    },
    "type": "string",
    "{customized property}": {}
  },
  "storeSettings": {
    "disableMetricsCollection": {},
    "maxConcurrentConnections": {},
    "type": "string"
    // For remaining properties, see StoreReadSettings objects
  }

For PaypalSource, use:

  "type": "PaypalSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For PhoenixSource, use:

  "type": "PhoenixSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For PostgreSqlSource, use:

  "type": "PostgreSqlSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For PostgreSqlV2Source, use:

  "type": "PostgreSqlV2Source",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For PrestoSource, use:

  "type": "PrestoSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For QuickBooksSource, use:

  "type": "QuickBooksSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For RelationalSource, use:

  "type": "RelationalSource",
  "additionalColumns": {},
  "query": {}

For ResponsysSource, use:

  "type": "ResponsysSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For RestSource, use:

  "type": "RestSource",
  "additionalColumns": {},
  "additionalHeaders": {},
  "httpRequestTimeout": {},
  "paginationRules": {},
  "requestBody": {},
  "requestInterval": {},
  "requestMethod": {}

For SalesforceMarketingCloudSource, use:

  "type": "SalesforceMarketingCloudSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For SalesforceServiceCloudSource, use:

  "type": "SalesforceServiceCloudSource",
  "additionalColumns": {},
  "query": {},
  "readBehavior": {}

For SalesforceServiceCloudV2Source, use:

  "type": "SalesforceServiceCloudV2Source",
  "additionalColumns": {},
  "includeDeletedObjects": {},
  "SOQLQuery": {}

For SalesforceSource, use:

  "type": "SalesforceSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {},
  "readBehavior": {}

For SalesforceV2Source, use:

  "type": "SalesforceV2Source",
  "additionalColumns": {},
  "includeDeletedObjects": {},
  "queryTimeout": {},
  "SOQLQuery": {}

For SapBwSource, use:

  "type": "SapBwSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For SapCloudForCustomerSource, use:

  "type": "SapCloudForCustomerSource",
  "additionalColumns": {},
  "httpRequestTimeout": {},
  "query": {},
  "queryTimeout": {}

For SapEccSource, use:

  "type": "SapEccSource",
  "additionalColumns": {},
  "httpRequestTimeout": {},
  "query": {},
  "queryTimeout": {}

For SapHanaSource, use:

  "type": "SapHanaSource",
  "additionalColumns": {},
  "packetSize": {},
  "partitionOption": {},
  "partitionSettings": {
    "partitionColumnName": {}
  },
  "query": {},
  "queryTimeout": {}

For SapOdpSource, use:

  "type": "SapOdpSource",
  "additionalColumns": {},
  "extractionMode": {},
  "projection": {},
  "queryTimeout": {},
  "selection": {},
  "subscriberProcess": {}

For SapOpenHubSource, use:

  "type": "SapOpenHubSource",
  "additionalColumns": {},
  "baseRequestId": {},
  "customRfcReadTableFunctionModule": {},
  "excludeLastRequest": {},
  "queryTimeout": {},
  "sapDataColumnDelimiter": {}

For SapTableSource, use:

  "type": "SapTableSource",
  "additionalColumns": {},
  "batchSize": {},
  "customRfcReadTableFunctionModule": {},
  "partitionOption": {},
  "partitionSettings": {
    "maxPartitionsNumber": {},
    "partitionColumnName": {},
    "partitionLowerBound": {},
    "partitionUpperBound": {}
  },
  "queryTimeout": {},
  "rfcTableFields": {},
  "rfcTableOptions": {},
  "rowCount": {},
  "rowSkips": {},
  "sapDataColumnDelimiter": {}

For ServiceNowSource, use:

  "type": "ServiceNowSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For ServiceNowV2Source, use:

  "type": "ServiceNowV2Source",
  "additionalColumns": {},
  "expression": {
    "operands": [
      {
      "operator": "string",
      "type": "string",
      "value": "string"
  },
  "queryTimeout": {}

For SharePointOnlineListSource, use:

  "type": "SharePointOnlineListSource",
  "httpRequestTimeout": {},
  "query": {}

For ShopifySource, use:

  "type": "ShopifySource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For SnowflakeSource, use:

  "type": "SnowflakeSource",
  "exportSettings": {
    "additionalCopyOptions": {
      "{customized property}": {}
    },
    "additionalFormatOptions": {
      "{customized property}": {}
    },
    "type": "string",
    "{customized property}": {}
  },
  "query": {}

For SnowflakeV2Source, use:

  "type": "SnowflakeV2Source",
  "exportSettings": {
    "additionalCopyOptions": {
      "{customized property}": {}
    },
    "additionalFormatOptions": {
      "{customized property}": {}
    },
    "type": "string",
    "{customized property}": {}
  },
  "query": {}

For SparkSource, use:

  "type": "SparkSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For SqlDWSource, use:

  "type": "SqlDWSource",
  "additionalColumns": {},
  "isolationLevel": {},
  "partitionOption": {},
  "partitionSettings": {
    "partitionColumnName": {},
    "partitionLowerBound": {},
    "partitionUpperBound": {}
  },
  "queryTimeout": {},
  "sqlReaderQuery": {},
  "sqlReaderStoredProcedureName": {},
  "storedProcedureParameters": {}

For SqlMISource, use:

  "type": "SqlMISource",
  "additionalColumns": {},
  "isolationLevel": {},
  "partitionOption": {},
  "partitionSettings": {
    "partitionColumnName": {},
    "partitionLowerBound": {},
    "partitionUpperBound": {}
  },
  "produceAdditionalTypes": {},
  "queryTimeout": {},
  "sqlReaderQuery": {},
  "sqlReaderStoredProcedureName": {},
  "storedProcedureParameters": {}

For SqlServerSource, use:

  "type": "SqlServerSource",
  "additionalColumns": {},
  "isolationLevel": {},
  "partitionOption": {},
  "partitionSettings": {
    "partitionColumnName": {},
    "partitionLowerBound": {},
    "partitionUpperBound": {}
  },
  "produceAdditionalTypes": {},
  "queryTimeout": {},
  "sqlReaderQuery": {},
  "sqlReaderStoredProcedureName": {},
  "storedProcedureParameters": {}

For SqlSource, use:

  "type": "SqlSource",
  "additionalColumns": {},
  "isolationLevel": {},
  "partitionOption": {},
  "partitionSettings": {
    "partitionColumnName": {},
    "partitionLowerBound": {},
    "partitionUpperBound": {}
  },
  "queryTimeout": {},
  "sqlReaderQuery": {},
  "sqlReaderStoredProcedureName": {},
  "storedProcedureParameters": {}

For SquareSource, use:

  "type": "SquareSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For SybaseSource, use:

  "type": "SybaseSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For TeradataSource, use:

  "type": "TeradataSource",
  "additionalColumns": {},
  "partitionOption": {},
  "partitionSettings": {
    "partitionColumnName": {},
    "partitionLowerBound": {},
    "partitionUpperBound": {}
  },
  "query": {},
  "queryTimeout": {}

For VerticaSource, use:

  "type": "VerticaSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For WarehouseSource, use:

  "type": "WarehouseSource",
  "additionalColumns": {},
  "isolationLevel": {},
  "partitionOption": {},
  "partitionSettings": {
    "partitionColumnName": {},
    "partitionLowerBound": {},
    "partitionUpperBound": {}
  },
  "queryTimeout": {},
  "sqlReaderQuery": {},
  "sqlReaderStoredProcedureName": {},
  "storedProcedureParameters": {}

For WebSource, use:

  "type": "WebSource",
  "additionalColumns": {}

For XeroSource, use:

  "type": "XeroSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

For XmlSource, use:

  "type": "XmlSource",
  "additionalColumns": {},
  "formatSettings": {
    "compressionProperties": {
      "type": "string"
      // For remaining properties, see CompressionReadSettings objects
    },
    "detectDataType": {},
    "namespacePrefixes": {},
    "namespaces": {},
    "type": "string",
    "validationMode": {},
    "{customized property}": {}
  },
  "storeSettings": {
    "disableMetricsCollection": {},
    "maxConcurrentConnections": {},
    "type": "string"
    // For remaining properties, see StoreReadSettings objects
  }

For ZohoSource, use:

  "type": "ZohoSource",
  "additionalColumns": {},
  "query": {},
  "queryTimeout": {}

StoreReadSettings objects

Set the type property to specify the type of object.

For AmazonS3CompatibleReadSettings, use:

  "type": "AmazonS3CompatibleReadSettings",
  "deleteFilesAfterCompletion": {},
  "enablePartitionDiscovery": {},
  "fileListPath": {},
  "modifiedDatetimeEnd": {},
  "modifiedDatetimeStart": {},
  "partitionRootPath": {},
  "prefix": {},
  "recursive": {},
  "wildcardFileName": {},
  "wildcardFolderPath": {}

For AmazonS3ReadSettings, use:

  "type": "AmazonS3ReadSettings",
  "deleteFilesAfterCompletion": {},
  "enablePartitionDiscovery": {},
  "fileListPath": {},
  "modifiedDatetimeEnd": {},
  "modifiedDatetimeStart": {},
  "partitionRootPath": {},
  "prefix": {},
  "recursive": {},
  "wildcardFileName": {},
  "wildcardFolderPath": {}

For AzureBlobFSReadSettings, use:

  "type": "AzureBlobFSReadSettings",
  "deleteFilesAfterCompletion": {},
  "enablePartitionDiscovery": {},
  "fileListPath": {},
  "modifiedDatetimeEnd": {},
  "modifiedDatetimeStart": {},
  "partitionRootPath": {},
  "recursive": {},
  "wildcardFileName": {},
  "wildcardFolderPath": {}

For AzureBlobStorageReadSettings, use:

  "type": "AzureBlobStorageReadSettings",
  "deleteFilesAfterCompletion": {},
  "enablePartitionDiscovery": {},
  "fileListPath": {},
  "modifiedDatetimeEnd": {},
  "modifiedDatetimeStart": {},
  "partitionRootPath": {},
  "prefix": {},
  "recursive": {},
  "wildcardFileName": {},
  "wildcardFolderPath": {}

For AzureDataLakeStoreReadSettings, use:

  "type": "AzureDataLakeStoreReadSettings",
  "deleteFilesAfterCompletion": {},
  "enablePartitionDiscovery": {},
  "fileListPath": {},
  "listAfter": {},
  "listBefore": {},
  "modifiedDatetimeEnd": {},
  "modifiedDatetimeStart": {},
  "partitionRootPath": {},
  "recursive": {},
  "wildcardFileName": {},
  "wildcardFolderPath": {}

For AzureFileStorageReadSettings, use:

  "type": "AzureFileStorageReadSettings",
  "deleteFilesAfterCompletion": {},
  "enablePartitionDiscovery": {},
  "fileListPath": {},
  "modifiedDatetimeEnd": {},
  "modifiedDatetimeStart": {},
  "partitionRootPath": {},
  "prefix": {},
  "recursive": {},
  "wildcardFileName": {},
  "wildcardFolderPath": {}

For FileServerReadSettings, use:

  "type": "FileServerReadSettings",
  "deleteFilesAfterCompletion": {},
  "enablePartitionDiscovery": {},
  "fileFilter": {},
  "fileListPath": {},
  "modifiedDatetimeEnd": {},
  "modifiedDatetimeStart": {},
  "partitionRootPath": {},
  "recursive": {},
  "wildcardFileName": {},
  "wildcardFolderPath": {}

For FtpReadSettings, use:

  "type": "FtpReadSettings",
  "deleteFilesAfterCompletion": {},
  "disableChunking": {},
  "enablePartitionDiscovery": {},
  "fileListPath": {},
  "partitionRootPath": {},
  "recursive": {},
  "useBinaryTransfer": {},
  "wildcardFileName": {},
  "wildcardFolderPath": {}

For GoogleCloudStorageReadSettings, use:

  "type": "GoogleCloudStorageReadSettings",
  "deleteFilesAfterCompletion": {},
  "enablePartitionDiscovery": {},
  "fileListPath": {},
  "modifiedDatetimeEnd": {},
  "modifiedDatetimeStart": {},
  "partitionRootPath": {},
  "prefix": {},
  "recursive": {},
  "wildcardFileName": {},
  "wildcardFolderPath": {}

For HdfsReadSettings, use:

  "type": "HdfsReadSettings",
  "deleteFilesAfterCompletion": {},
  "distcpSettings": {
    "distcpOptions": {},
    "resourceManagerEndpoint": {},
    "tempScriptPath": {}
  },
  "enablePartitionDiscovery": {},
  "fileListPath": {},
  "modifiedDatetimeEnd": {},
  "modifiedDatetimeStart": {},
  "partitionRootPath": {},
  "recursive": {},
  "wildcardFileName": {},
  "wildcardFolderPath": {}

For HttpReadSettings, use:

  "type": "HttpReadSettings",
  "additionalColumns": {},
  "additionalHeaders": {},
  "requestBody": {},
  "requestMethod": {},
  "requestTimeout": {}

For LakeHouseReadSettings, use:

  "type": "LakeHouseReadSettings",
  "deleteFilesAfterCompletion": {},
  "enablePartitionDiscovery": {},
  "fileListPath": {},
  "modifiedDatetimeEnd": {},
  "modifiedDatetimeStart": {},
  "partitionRootPath": {},
  "recursive": {},
  "wildcardFileName": {},
  "wildcardFolderPath": {}

For OracleCloudStorageReadSettings, use:

  "type": "OracleCloudStorageReadSettings",
  "deleteFilesAfterCompletion": {},
  "enablePartitionDiscovery": {},
  "fileListPath": {},
  "modifiedDatetimeEnd": {},
  "modifiedDatetimeStart": {},
  "partitionRootPath": {},
  "prefix": {},
  "recursive": {},
  "wildcardFileName": {},
  "wildcardFolderPath": {}

For SftpReadSettings, use:

  "type": "SftpReadSettings",
  "deleteFilesAfterCompletion": {},
  "disableChunking": {},
  "enablePartitionDiscovery": {},
  "fileListPath": {},
  "modifiedDatetimeEnd": {},
  "modifiedDatetimeStart": {},
  "partitionRootPath": {},
  "recursive": {},
  "wildcardFileName": {},
  "wildcardFolderPath": {}

CompressionReadSettings objects

Set the type property to specify the type of object.

For TarGZipReadSettings, use:

  "type": "TarGZipReadSettings",
  "preserveCompressionFileNameAsFolder": {}

For TarReadSettings, use:

  "type": "TarReadSettings",
  "preserveCompressionFileNameAsFolder": {}

For ZipDeflateReadSettings, use:

  "type": "ZipDeflateReadSettings",
  "preserveZipFileNameAsFolder": {}

SecretBase objects

Set the type property to specify the type of object.

For AzureKeyVaultSecret, use:

  "type": "AzureKeyVaultSecret",
  "secretName": {},
  "secretVersion": {},
  "store": {
    "parameters": {
      "{customized property}": {}
    },
    "referenceName": "string",
    "type": "LinkedServiceReference"
  }

For SecureString, use:

  "type": "SecureString",
  "value": "string"

FormatReadSettings objects

Set the type property to specify the type of object.

For BinaryReadSettings, use:

  "type": "BinaryReadSettings",
  "compressionProperties": {
    "type": "string"
    // For remaining properties, see CompressionReadSettings objects
  }

For DelimitedTextReadSettings, use:

  "type": "DelimitedTextReadSettings",
  "compressionProperties": {
    "type": "string"
    // For remaining properties, see CompressionReadSettings objects
  },
  "skipLineCount": {}

For JsonReadSettings, use:

  "type": "JsonReadSettings",
  "compressionProperties": {
    "type": "string"
    // For remaining properties, see CompressionReadSettings objects
  }

For ParquetReadSettings, use:

  "type": "ParquetReadSettings",
  "compressionProperties": {
    "type": "string"
    // For remaining properties, see CompressionReadSettings objects
  }

For XmlReadSettings, use:

  "type": "XmlReadSettings",
  "compressionProperties": {
    "type": "string"
    // For remaining properties, see CompressionReadSettings objects
  },
  "detectDataType": {},
  "namespacePrefixes": {},
  "namespaces": {},
  "validationMode": {}

Property values

factories/pipelines

Name Description Value
type The resource type 'Microsoft.DataFactory/factories/pipelines'
apiVersion The resource api version '2018-06-01'
name The resource name

See how to set names and types for child resources in JSON ARM templates.
string (required)

Character limit: 1-260

Valid characters:
Can't use:
<>*#.%&:\\+?/ or control characters

Start with alphanumeric.
properties Properties of the pipeline. Pipeline (required)

Pipeline

Name Description Value
activities List of activities in pipeline. Activity[]
annotations List of tags that can be used for describing the Pipeline. any[]
concurrency The max number of concurrent runs for the pipeline. int

Constraints:
Min value = 1
description The description of the pipeline. string
folder The folder that this Pipeline is in. If not specified, Pipeline will appear at the root level. PipelineFolder
parameters List of parameters for pipeline. ParameterDefinitionSpecification
policy Pipeline Policy. PipelinePolicy
runDimensions Dimensions emitted by Pipeline. PipelineRunDimensions
variables List of variables for pipeline. VariableDefinitionSpecification

Activity

Name Description Value
dependsOn Activity depends on condition. ActivityDependency[]
description Activity description. string
name Activity name. string (required)
onInactiveMarkAs Status result of the activity when the state is set to Inactive. This is an optional property and if not provided when the activity is inactive, the status will be Succeeded by default. 'Failed'
'Skipped'
'Succeeded'
state Activity state. This is an optional property and if not provided, the state will be Active by default. 'Active'
'Inactive'
userProperties Activity user properties. UserProperty[]
type Set the object type AppendVariable
AzureDataExplorerCommand
AzureFunctionActivity
AzureMLBatchExecution
AzureMLExecutePipeline
AzureMLUpdateResource
Copy
Custom
DatabricksNotebook
DatabricksSparkJar
DatabricksSparkPython
DataLakeAnalyticsU-SQL
Delete
ExecuteDataFlow
ExecutePipeline
ExecuteSSISPackage
ExecuteWranglingDataflow
Fail
Filter
ForEach
GetMetadata
HDInsightHive
HDInsightMapReduce
HDInsightPig
HDInsightSpark
HDInsightStreaming
IfCondition
Lookup
Script
SetVariable
SparkJob
SqlServerStoredProcedure
Switch
SynapseNotebook
Until
Validation
Wait
WebActivity
WebHook (required)

ActivityDependency

Name Description Value
activity Activity name. string (required)
dependencyConditions Match-Condition for the dependency. String array containing any of:
'Completed'
'Failed'
'Skipped'
'Succeeded' (required)
{customized property}

UserProperty

Name Description Value
name User property name. string (required)
value User property value. Type: string (or Expression with resultType string).

AppendVariableActivity

Name Description Value
type Type of activity. 'AppendVariable' (required)
typeProperties Append Variable activity properties. AppendVariableActivityTypeProperties (required)

AppendVariableActivityTypeProperties

Name Description Value
value Value to be appended. Type: could be a static value matching type of the variable item or Expression with resultType matching type of the variable item
variableName Name of the variable whose value needs to be appended to. string

AzureDataExplorerCommandActivity

Name Description Value
type Type of activity. 'AzureDataExplorerCommand' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Azure Data Explorer command activity properties. AzureDataExplorerCommandActivityTypeProperties (required)

LinkedServiceReference

Name Description Value
parameters Arguments for LinkedService. ParameterValueSpecification
referenceName Reference LinkedService name. string (required)
type Linked service reference type. 'LinkedServiceReference' (required)

ParameterValueSpecification

Name Description Value
{customized property}

ActivityPolicy

Name Description Value
retry Maximum ordinary retry attempts. Default is 0. Type: integer (or Expression with resultType integer), minimum: 0.
retryIntervalInSeconds Interval between each retry attempt (in seconds). The default is 30 sec. int

Constraints:
Min value = 30
Max value = 86400
secureInput When set to true, Input from activity is considered as secure and will not be logged to monitoring. bool
secureOutput When set to true, Output from activity is considered as secure and will not be logged to monitoring. bool
timeout Specifies the timeout for the activity to run. The default timeout is 7 days. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
{customized property}

AzureDataExplorerCommandActivityTypeProperties

Name Description Value
command A control command, according to the Azure Data Explorer command syntax. Type: string (or Expression with resultType string).
commandTimeout Control command timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9]))..)

AzureFunctionActivity

Name Description Value
type Type of activity. 'AzureFunctionActivity' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Azure Function activity properties. AzureFunctionActivityTypeProperties (required)

AzureFunctionActivityTypeProperties

Name Description Value
body Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string).
functionName Name of the Function that the Azure Function Activity will call. Type: string (or Expression with resultType string)
headers Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). AzureFunctionActivityTypePropertiesHeaders
method Rest API method for target endpoint. 'DELETE'
'GET'
'HEAD'
'OPTIONS'
'POST'
'PUT'
'TRACE' (required)

AzureFunctionActivityTypePropertiesHeaders

Name Description Value
{customized property} string

AzureMLBatchExecutionActivity

Name Description Value
type Type of activity. 'AzureMLBatchExecution' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Azure ML Batch Execution activity properties. AzureMLBatchExecutionActivityTypeProperties (required)

AzureMLBatchExecutionActivityTypeProperties

Name Description Value
globalParameters Key,Value pairs to be passed to the Azure ML Batch Execution Service endpoint. Keys must match the names of web service parameters defined in the published Azure ML web service. Values will be passed in the GlobalParameters property of the Azure ML batch execution request. AzureMLBatchExecutionActivityTypePropertiesGlobalPar...
webServiceInputs Key,Value pairs, mapping the names of Azure ML endpoint's Web Service Inputs to AzureMLWebServiceFile objects specifying the input Blob locations.. This information will be passed in the WebServiceInputs property of the Azure ML batch execution request. AzureMLBatchExecutionActivityTypePropertiesWebServic...
webServiceOutputs Key,Value pairs, mapping the names of Azure ML endpoint's Web Service Outputs to AzureMLWebServiceFile objects specifying the output Blob locations. This information will be passed in the WebServiceOutputs property of the Azure ML batch execution request. AzureMLBatchExecutionActivityTypePropertiesWebServic...

AzureMLBatchExecutionActivityTypePropertiesGlobalPar...

Name Description Value
{customized property}

AzureMLBatchExecutionActivityTypePropertiesWebServic...

Name Description Value
{customized property} AzureMLWebServiceFile

AzureMLWebServiceFile

Name Description Value
filePath The relative file path, including container name, in the Azure Blob Storage specified by the LinkedService. Type: string (or Expression with resultType string).
linkedServiceName Reference to an Azure Storage LinkedService, where Azure ML WebService Input/Output file located. LinkedServiceReference (required)

AzureMLBatchExecutionActivityTypePropertiesWebServic...

Name Description Value
{customized property} AzureMLWebServiceFile

AzureMLExecutePipelineActivity

Name Description Value
type Type of activity. 'AzureMLExecutePipeline' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Azure ML Execute Pipeline activity properties. AzureMLExecutePipelineActivityTypeProperties (required)

AzureMLExecutePipelineActivityTypeProperties

Name Description Value
continueOnStepFailure Whether to continue execution of other steps in the PipelineRun if a step fails. This information will be passed in the continueOnStepFailure property of the published pipeline execution request. Type: boolean (or Expression with resultType boolean).
dataPathAssignments Dictionary used for changing data path assignments without retraining. Values will be passed in the dataPathAssignments property of the published pipeline execution request. Type: object (or Expression with resultType object).
experimentName Run history experiment name of the pipeline run. This information will be passed in the ExperimentName property of the published pipeline execution request. Type: string (or Expression with resultType string).
mlParentRunId The parent Azure ML Service pipeline run id. This information will be passed in the ParentRunId property of the published pipeline execution request. Type: string (or Expression with resultType string).
mlPipelineEndpointId ID of the published Azure ML pipeline endpoint. Type: string (or Expression with resultType string).
mlPipelineId ID of the published Azure ML pipeline. Type: string (or Expression with resultType string).
mlPipelineParameters Key,Value pairs to be passed to the published Azure ML pipeline endpoint. Keys must match the names of pipeline parameters defined in the published pipeline. Values will be passed in the ParameterAssignments property of the published pipeline execution request. Type: object with key value pairs (or Expression with resultType object).
version Version of the published Azure ML pipeline endpoint. Type: string (or Expression with resultType string).

AzureMLUpdateResourceActivity

Name Description Value
type Type of activity. 'AzureMLUpdateResource' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Azure ML Update Resource management activity properties. AzureMLUpdateResourceActivityTypeProperties (required)

AzureMLUpdateResourceActivityTypeProperties

Name Description Value
trainedModelFilePath The relative file path in trainedModelLinkedService to represent the .ilearner file that will be uploaded by the update operation. Type: string (or Expression with resultType string).
trainedModelLinkedServiceName Name of Azure Storage linked service holding the .ilearner file that will be uploaded by the update operation. LinkedServiceReference (required)
trainedModelName Name of the Trained Model module in the Web Service experiment to be updated. Type: string (or Expression with resultType string).

CopyActivity

Name Description Value
type Type of activity. 'Copy' (required)
inputs List of inputs for the activity. DatasetReference[]
linkedServiceName Linked service reference. LinkedServiceReference
outputs List of outputs for the activity. DatasetReference[]
policy Activity policy. ActivityPolicy
typeProperties Copy activity properties. CopyActivityTypeProperties (required)

DatasetReference

Name Description Value
parameters Arguments for dataset. ParameterValueSpecification
referenceName Reference dataset name. string (required)
type Dataset reference type. 'DatasetReference' (required)

CopyActivityTypeProperties

Name Description Value
dataIntegrationUnits Maximum number of data integration units that can be used to perform this data movement. Type: integer (or Expression with resultType integer), minimum: 0.
enableSkipIncompatibleRow Whether to skip incompatible row. Default value is false. Type: boolean (or Expression with resultType boolean).
enableStaging Specifies whether to copy data via an interim staging. Default value is false. Type: boolean (or Expression with resultType boolean).
logSettings Log settings customer needs provide when enabling log. LogSettings
logStorageSettings (Deprecated. Please use LogSettings) Log storage settings customer need to provide when enabling session log. LogStorageSettings
parallelCopies Maximum number of concurrent sessions opened on the source or sink to avoid overloading the data store. Type: integer (or Expression with resultType integer), minimum: 0.
preserve Preserve rules. any[]
preserveRules Preserve Rules. any[]
redirectIncompatibleRowSettings Redirect incompatible row settings when EnableSkipIncompatibleRow is true. RedirectIncompatibleRowSettings
sink Copy activity sink. CopySink (required)
skipErrorFile Specify the fault tolerance for data consistency. SkipErrorFile
source Copy activity source. CopySource (required)
stagingSettings Specifies interim staging settings when EnableStaging is true. StagingSettings
translator Copy activity translator. If not specified, tabular translator is used.
validateDataConsistency Whether to enable Data Consistency validation. Type: boolean (or Expression with resultType boolean).

LogSettings

Name Description Value
copyActivityLogSettings Specifies settings for copy activity log. CopyActivityLogSettings
enableCopyActivityLog Specifies whether to enable copy activity log. Type: boolean (or Expression with resultType boolean).
logLocationSettings Log location settings customer needs to provide when enabling log. LogLocationSettings (required)

CopyActivityLogSettings

Name Description Value
enableReliableLogging Specifies whether to enable reliable logging. Type: boolean (or Expression with resultType boolean).
logLevel Gets or sets the log level, support: Info, Warning. Type: string (or Expression with resultType string).

LogLocationSettings

Name Description Value
linkedServiceName Log storage linked service reference. LinkedServiceReference (required)
path The path to storage for storing detailed logs of activity execution. Type: string (or Expression with resultType string).

LogStorageSettings

Name Description Value
enableReliableLogging Specifies whether to enable reliable logging. Type: boolean (or Expression with resultType boolean).
linkedServiceName Log storage linked service reference. LinkedServiceReference (required)
logLevel Gets or sets the log level, support: Info, Warning. Type: string (or Expression with resultType string).
path The path to storage for storing detailed logs of activity execution. Type: string (or Expression with resultType string).
{customized property}

RedirectIncompatibleRowSettings

Name Description Value
linkedServiceName Name of the Azure Storage, Storage SAS, or Azure Data Lake Store linked service used for redirecting incompatible row. Must be specified if redirectIncompatibleRowSettings is specified. Type: string (or Expression with resultType string).
path The path for storing the redirect incompatible row data. Type: string (or Expression with resultType string).
{customized property}

CopySink

Name Description Value
disableMetricsCollection If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean).
maxConcurrentConnections The maximum concurrent connection count for the sink data store. Type: integer (or Expression with resultType integer).
sinkRetryCount Sink retry count. Type: integer (or Expression with resultType integer).
sinkRetryWait Sink retry wait. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
writeBatchSize Write batch size. Type: integer (or Expression with resultType integer), minimum: 0.
writeBatchTimeout Write batch timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
type Set the object type AvroSink
AzureBlobFSSink
AzureDatabricksDeltaLakeSink
AzureDataExplorerSink
AzureDataLakeStoreSink
AzureMySqlSink
AzurePostgreSqlSink
AzureQueueSink
AzureSearchIndexSink
AzureSqlSink
AzureTableSink
BinarySink
BlobSink
CommonDataServiceForAppsSink
CosmosDbMongoDbApiSink
CosmosDbSqlApiSink
DelimitedTextSink
DocumentDbCollectionSink
DynamicsCrmSink
DynamicsSink
FileSystemSink
InformixSink
JsonSink
LakeHouseTableSink
MicrosoftAccessSink
MongoDbAtlasSink
MongoDbV2Sink
OdbcSink
OracleSink
OrcSink
ParquetSink
RestSink
SalesforceServiceCloudSink
SalesforceServiceCloudV2Sink
SalesforceSink
SalesforceV2Sink
SapCloudForCustomerSink
SnowflakeSink
SnowflakeV2Sink
SqlDWSink
SqlMISink
SqlServerSink
SqlSink
WarehouseSink (required)

AvroSink

Name Description Value
type Copy sink type. 'AvroSink' (required)
formatSettings Avro format settings. AvroWriteSettings
storeSettings Avro store settings. StoreWriteSettings

AvroWriteSettings

Name Description Value
fileNamePrefix Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string).
maxRowsPerFile Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer).
recordName Top level record name in write result, which is required in AVRO spec. string
recordNamespace Record namespace in the write result. string
type The write setting type. string (required)
{customized property}

StoreWriteSettings

Name Description Value
copyBehavior The type of copy behavior for copy sink.
disableMetricsCollection If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean).
maxConcurrentConnections The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer).
metadata Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). MetadataItem[]
type Set the object type AzureBlobFSWriteSettings
AzureBlobStorageWriteSettings
AzureDataLakeStoreWriteSettings
AzureFileStorageWriteSettings
FileServerWriteSettings
LakeHouseWriteSettings
SftpWriteSettings (required)

MetadataItem

Name Description Value
name Metadata item key name. Type: string (or Expression with resultType string).
value Metadata item value. Type: string (or Expression with resultType string).

AzureBlobFSWriteSettings

Name Description Value
type The write setting type. 'AzureBlobFSWriteSettings' (required)
blockSizeInMB Indicates the block size(MB) when writing data to blob. Type: integer (or Expression with resultType integer).

AzureBlobStorageWriteSettings

Name Description Value
type The write setting type. 'AzureBlobStorageWriteSettings' (required)
blockSizeInMB Indicates the block size(MB) when writing data to blob. Type: integer (or Expression with resultType integer).

AzureDataLakeStoreWriteSettings

Name Description Value
type The write setting type. 'AzureDataLakeStoreWriteSettings' (required)
expiryDateTime Specifies the expiry time of the written files. The time is applied to the UTC time zone in the format of "2018-12-01T05:00:00Z". Default value is NULL. Type: string (or Expression with resultType string).

AzureFileStorageWriteSettings

Name Description Value
type The write setting type. 'AzureFileStorageWriteSettings' (required)

FileServerWriteSettings

Name Description Value
type The write setting type. 'FileServerWriteSettings' (required)

LakeHouseWriteSettings

Name Description Value
type The write setting type. 'LakeHouseWriteSettings' (required)

SftpWriteSettings

Name Description Value
type The write setting type. 'SftpWriteSettings' (required)
operationTimeout Specifies the timeout for writing each chunk to SFTP server. Default value: 01:00:00 (one hour). Type: string (or Expression with resultType string).
useTempFileRename Upload to temporary file(s) and rename. Disable this option if your SFTP server doesn't support rename operation. Type: boolean (or Expression with resultType boolean).

AzureBlobFSSink

Name Description Value
type Copy sink type. 'AzureBlobFSSink' (required)
copyBehavior The type of copy behavior for copy sink. Type: string (or Expression with resultType string).
metadata Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). MetadataItem[]

AzureDatabricksDeltaLakeSink

Name Description Value
type Copy sink type. 'AzureDatabricksDeltaLakeSink' (required)
importSettings Azure Databricks Delta Lake import settings. AzureDatabricksDeltaLakeImportCommand
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).

AzureDatabricksDeltaLakeImportCommand

Name Description Value
dateFormat Specify the date format for csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string).
timestampFormat Specify the timestamp format for csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string).
type The import setting type. string (required)
{customized property}

AzureDataExplorerSink

Name Description Value
type Copy sink type. 'AzureDataExplorerSink' (required)
flushImmediately If set to true, any aggregation will be skipped. Default is false. Type: boolean.
ingestionMappingAsJson An explicit column mapping description provided in a json format. Type: string.
ingestionMappingName A name of a pre-created csv mapping that was defined on the target Kusto table. Type: string.

AzureDataLakeStoreSink

Name Description Value
type Copy sink type. 'AzureDataLakeStoreSink' (required)
copyBehavior The type of copy behavior for copy sink. Type: string (or Expression with resultType string).
enableAdlsSingleFileParallel Single File Parallel.

AzureMySqlSink

Name Description Value
type Copy sink type. 'AzureMySqlSink' (required)
preCopyScript A query to execute before starting the copy. Type: string (or Expression with resultType string).

AzurePostgreSqlSink

Name Description Value
type Copy sink type. 'AzurePostgreSqlSink' (required)
preCopyScript A query to execute before starting the copy. Type: string (or Expression with resultType string).

AzureQueueSink

Name Description Value
type Copy sink type. 'AzureQueueSink' (required)

AzureSearchIndexSink

Name Description Value
type Copy sink type. 'AzureSearchIndexSink' (required)
writeBehavior Specify the write behavior when upserting documents into Azure Search Index. 'Merge'
'Upload'

AzureSqlSink

Name Description Value
type Copy sink type. 'AzureSqlSink' (required)
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).
sqlWriterStoredProcedureName SQL writer stored procedure name. Type: string (or Expression with resultType string).
sqlWriterTableType SQL writer table type. Type: string (or Expression with resultType string).
sqlWriterUseTableLock Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean).
storedProcedureParameters SQL stored procedure parameters.
storedProcedureTableTypeParameterName The stored procedure parameter name of the table type. Type: string (or Expression with resultType string).
tableOption The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string).
upsertSettings SQL upsert settings. SqlUpsertSettings
writeBehavior Write behavior when copying data into Azure SQL. Type: SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum)

SqlUpsertSettings

Name Description Value
interimSchemaName Schema name for interim table. Type: string (or Expression with resultType string).
keys Key column names for unique row identification. Type: array of strings (or Expression with resultType array of strings).
useTempDB Specifies whether to use temp db for upsert interim table. Type: boolean (or Expression with resultType boolean).

AzureTableSink

Name Description Value
type Copy sink type. 'AzureTableSink' (required)
azureTableDefaultPartitionKeyValue Azure Table default partition key value. Type: string (or Expression with resultType string).
azureTableInsertType Azure Table insert type. Type: string (or Expression with resultType string).
azureTablePartitionKeyName Azure Table partition key name. Type: string (or Expression with resultType string).
azureTableRowKeyName Azure Table row key name. Type: string (or Expression with resultType string).

BinarySink

Name Description Value
type Copy sink type. 'BinarySink' (required)
storeSettings Binary store settings. StoreWriteSettings

BlobSink

Name Description Value
type Copy sink type. 'BlobSink' (required)
blobWriterAddHeader Blob writer add header. Type: boolean (or Expression with resultType boolean).
blobWriterDateTimeFormat Blob writer date time format. Type: string (or Expression with resultType string).
blobWriterOverwriteFiles Blob writer overwrite files. Type: boolean (or Expression with resultType boolean).
copyBehavior The type of copy behavior for copy sink.
metadata Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). MetadataItem[]

CommonDataServiceForAppsSink

Name Description Value
type Copy sink type. 'CommonDataServiceForAppsSink' (required)
alternateKeyName The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string).
ignoreNullValues The flag indicating whether to ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean).
writeBehavior The write behavior for the operation. 'Upsert' (required)

CosmosDbMongoDbApiSink

Name Description Value
type Copy sink type. 'CosmosDbMongoDbApiSink' (required)
writeBehavior Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string).

CosmosDbSqlApiSink

Name Description Value
type Copy sink type. 'CosmosDbSqlApiSink' (required)
writeBehavior Describes how to write data to Azure Cosmos DB. Type: string (or Expression with resultType string). Allowed values: insert and upsert.

DelimitedTextSink

Name Description Value
type Copy sink type. 'DelimitedTextSink' (required)
formatSettings DelimitedText format settings. DelimitedTextWriteSettings
storeSettings DelimitedText store settings. StoreWriteSettings

DelimitedTextWriteSettings

Name Description Value
fileExtension The file extension used to create the files. Type: string (or Expression with resultType string).
fileNamePrefix Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string).
maxRowsPerFile Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer).
quoteAllText Indicates whether string values should always be enclosed with quotes. Type: boolean (or Expression with resultType boolean).
type The write setting type. string (required)
{customized property}

DocumentDbCollectionSink

Name Description Value
type Copy sink type. 'DocumentDbCollectionSink' (required)
nestingSeparator Nested properties separator. Default is . (dot). Type: string (or Expression with resultType string).
writeBehavior Describes how to write data to Azure Cosmos DB. Type: string (or Expression with resultType string). Allowed values: insert and upsert.

DynamicsCrmSink

Name Description Value
type Copy sink type. 'DynamicsCrmSink' (required)
alternateKeyName The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string).
ignoreNullValues The flag indicating whether to ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean).
writeBehavior The write behavior for the operation. 'Upsert' (required)

DynamicsSink

Name Description Value
type Copy sink type. 'DynamicsSink' (required)
alternateKeyName The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string).
ignoreNullValues The flag indicating whether ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean).
writeBehavior The write behavior for the operation. 'Upsert' (required)

FileSystemSink

Name Description Value
type Copy sink type. 'FileSystemSink' (required)
copyBehavior The type of copy behavior for copy sink.

InformixSink

Name Description Value
type Copy sink type. 'InformixSink' (required)
preCopyScript A query to execute before starting the copy. Type: string (or Expression with resultType string).

JsonSink

Name Description Value
type Copy sink type. 'JsonSink' (required)
formatSettings Json format settings. JsonWriteSettings
storeSettings Json store settings. StoreWriteSettings

JsonWriteSettings

Name Description Value
filePattern File pattern of JSON. This setting controls the way a collection of JSON objects will be treated. The default value is 'setOfObjects'. It is case-sensitive.
type The write setting type. string (required)
{customized property}

LakeHouseTableSink

Name Description Value
type Copy sink type. 'LakeHouseTableSink' (required)
partitionNameList Specify the partition column names from sink columns. Type: array of objects (or Expression with resultType array of objects).
partitionOption Create partitions in folder structure based on one or multiple columns. Each distinct column value (pair) will be a new partition. Possible values include: "None", "PartitionByKey".
tableActionOption The type of table action for LakeHouse Table sink. Possible values include: "None", "Append", "Overwrite".

MicrosoftAccessSink

Name Description Value
type Copy sink type. 'MicrosoftAccessSink' (required)
preCopyScript A query to execute before starting the copy. Type: string (or Expression with resultType string).

MongoDbAtlasSink

Name Description Value
type Copy sink type. 'MongoDbAtlasSink' (required)
writeBehavior Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string).

MongoDbV2Sink

Name Description Value
type Copy sink type. 'MongoDbV2Sink' (required)
writeBehavior Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string).

OdbcSink

Name Description Value
type Copy sink type. 'OdbcSink' (required)
preCopyScript A query to execute before starting the copy. Type: string (or Expression with resultType string).

OracleSink

Name Description Value
type Copy sink type. 'OracleSink' (required)
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).

OrcSink

Name Description Value
type Copy sink type. 'OrcSink' (required)
formatSettings ORC format settings. OrcWriteSettings
storeSettings ORC store settings. StoreWriteSettings

OrcWriteSettings

Name Description Value
fileNamePrefix Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string).
maxRowsPerFile Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer).
type The write setting type. string (required)
{customized property}

ParquetSink

Name Description Value
type Copy sink type. 'ParquetSink' (required)
formatSettings Parquet format settings. ParquetWriteSettings
storeSettings Parquet store settings. StoreWriteSettings

ParquetWriteSettings

Name Description Value
fileNamePrefix Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string).
maxRowsPerFile Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer).
type The write setting type. string (required)
{customized property}

RestSink

Name Description Value
type Copy sink type. 'RestSink' (required)
additionalHeaders The additional HTTP headers in the request to the RESTful API. Type: key value pairs (value should be string type).
httpCompressionType Http Compression Type to Send data in compressed format with Optimal Compression Level, Default is None. And The Only Supported option is Gzip. Type: string (or Expression with resultType string).
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
requestInterval The time to await before sending next request, in milliseconds
requestMethod The HTTP method used to call the RESTful API. The default is POST. Type: string (or Expression with resultType string).

SalesforceServiceCloudSink

Name Description Value
type Copy sink type. 'SalesforceServiceCloudSink' (required)
externalIdFieldName The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string).
ignoreNullValues The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean).
writeBehavior The write behavior for the operation. Default is Insert. 'Insert'
'Upsert'

SalesforceServiceCloudV2Sink

Name Description Value
type Copy sink type. 'SalesforceServiceCloudV2Sink' (required)
externalIdFieldName The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string).
ignoreNullValues The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean).
writeBehavior The write behavior for the operation. Default is Insert. 'Insert'
'Upsert'

SalesforceSink

Name Description Value
type Copy sink type. 'SalesforceSink' (required)
externalIdFieldName The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string).
ignoreNullValues The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean).
writeBehavior The write behavior for the operation. Default is Insert. 'Insert'
'Upsert'

SalesforceV2Sink

Name Description Value
type Copy sink type. 'SalesforceV2Sink' (required)
externalIdFieldName The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string).
ignoreNullValues The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean).
writeBehavior The write behavior for the operation. Default is Insert. 'Insert'
'Upsert'

SapCloudForCustomerSink

Name Description Value
type Copy sink type. 'SapCloudForCustomerSink' (required)
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
writeBehavior The write behavior for the operation. Default is 'Insert'. 'Insert'
'Update'

SnowflakeSink

Name Description Value
type Copy sink type. 'SnowflakeSink' (required)
importSettings Snowflake import settings. SnowflakeImportCopyCommand
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).

SnowflakeImportCopyCommand

Name Description Value
additionalCopyOptions Additional copy options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": "'HH24:MI:SS.FF'" } SnowflakeImportCopyCommandAdditionalCopyOptions
additionalFormatOptions Additional format options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalFormatOptions": { "FORCE": "TRUE", "LOAD_UNCERTAIN_FILES": "'FALSE'" } SnowflakeImportCopyCommandAdditionalFormatOptions
type The import setting type. string (required)
{customized property}

SnowflakeImportCopyCommandAdditionalCopyOptions

Name Description Value
{customized property}

SnowflakeImportCopyCommandAdditionalFormatOptions

Name Description Value
{customized property}

SnowflakeV2Sink

Name Description Value
type Copy sink type. 'SnowflakeV2Sink' (required)
importSettings Snowflake import settings. SnowflakeImportCopyCommand
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).

SqlDWSink

Name Description Value
type Copy sink type. 'SqlDWSink' (required)
allowCopyCommand Indicates to use Copy Command to copy data into SQL Data Warehouse. Type: boolean (or Expression with resultType boolean).
allowPolyBase Indicates to use PolyBase to copy data into SQL Data Warehouse when applicable. Type: boolean (or Expression with resultType boolean).
copyCommandSettings Specifies Copy Command related settings when allowCopyCommand is true. DWCopyCommandSettings
polyBaseSettings Specifies PolyBase-related settings when allowPolyBase is true. PolybaseSettings
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).
sqlWriterUseTableLock Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean).
tableOption The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string).
upsertSettings SQL DW upsert settings. SqlDWUpsertSettings
writeBehavior Write behavior when copying data into azure SQL DW. Type: SqlDWWriteBehaviorEnum (or Expression with resultType SqlDWWriteBehaviorEnum)

DWCopyCommandSettings

Name Description Value
additionalOptions Additional options directly passed to SQL DW in Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalOptions": { "MAXERRORS": "1000", "DATEFORMAT": "'ymd'" } DWCopyCommandSettingsAdditionalOptions
defaultValues Specifies the default values for each target column in SQL DW. The default values in the property overwrite the DEFAULT constraint set in the DB, and identity column cannot have a default value. Type: array of objects (or Expression with resultType array of objects). DWCopyCommandDefaultValue[]

DWCopyCommandSettingsAdditionalOptions

Name Description Value
{customized property} string

DWCopyCommandDefaultValue

Name Description Value
columnName Column name. Type: object (or Expression with resultType string).
defaultValue The default value of the column. Type: object (or Expression with resultType string).

PolybaseSettings

Name Description Value
rejectSampleValue Determines the number of rows to attempt to retrieve before the PolyBase recalculates the percentage of rejected rows. Type: integer (or Expression with resultType integer), minimum: 0.
rejectType Reject type. 'percentage'
'value'
rejectValue Specifies the value or the percentage of rows that can be rejected before the query fails. Type: number (or Expression with resultType number), minimum: 0.
useTypeDefault Specifies how to handle missing values in delimited text files when PolyBase retrieves data from the text file. Type: boolean (or Expression with resultType boolean).
{customized property}

SqlDWUpsertSettings

Name Description Value
interimSchemaName Schema name for interim table. Type: string (or Expression with resultType string).
keys Key column names for unique row identification. Type: array of strings (or Expression with resultType array of strings).

SqlMISink

Name Description Value
type Copy sink type. 'SqlMISink' (required)
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).
sqlWriterStoredProcedureName SQL writer stored procedure name. Type: string (or Expression with resultType string).
sqlWriterTableType SQL writer table type. Type: string (or Expression with resultType string).
sqlWriterUseTableLock Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean).
storedProcedureParameters SQL stored procedure parameters.
storedProcedureTableTypeParameterName The stored procedure parameter name of the table type. Type: string (or Expression with resultType string).
tableOption The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string).
upsertSettings SQL upsert settings. SqlUpsertSettings
writeBehavior White behavior when copying data into azure SQL MI. Type: string (or Expression with resultType string)

SqlServerSink

Name Description Value
type Copy sink type. 'SqlServerSink' (required)
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).
sqlWriterStoredProcedureName SQL writer stored procedure name. Type: string (or Expression with resultType string).
sqlWriterTableType SQL writer table type. Type: string (or Expression with resultType string).
sqlWriterUseTableLock Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean).
storedProcedureParameters SQL stored procedure parameters.
storedProcedureTableTypeParameterName The stored procedure parameter name of the table type. Type: string (or Expression with resultType string).
tableOption The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string).
upsertSettings SQL upsert settings. SqlUpsertSettings
writeBehavior Write behavior when copying data into sql server. Type: string (or Expression with resultType string).

SqlSink

Name Description Value
type Copy sink type. 'SqlSink' (required)
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).
sqlWriterStoredProcedureName SQL writer stored procedure name. Type: string (or Expression with resultType string).
sqlWriterTableType SQL writer table type. Type: string (or Expression with resultType string).
sqlWriterUseTableLock Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean).
storedProcedureParameters SQL stored procedure parameters.
storedProcedureTableTypeParameterName The stored procedure parameter name of the table type. Type: string (or Expression with resultType string).
tableOption The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string).
upsertSettings SQL upsert settings. SqlUpsertSettings
writeBehavior Write behavior when copying data into sql. Type: string (or Expression with resultType string).

WarehouseSink

Name Description Value
type Copy sink type. 'WarehouseSink' (required)
allowCopyCommand Indicates to use Copy Command to copy data into SQL Data Warehouse. Type: boolean (or Expression with resultType boolean).
copyCommandSettings Specifies Copy Command related settings when allowCopyCommand is true. DWCopyCommandSettings
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).
tableOption The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string).
writeBehavior Write behavior when copying data into azure Microsoft Fabric Data Warehouse. Type: DWWriteBehaviorEnum (or Expression with resultType DWWriteBehaviorEnum)

SkipErrorFile

Name Description Value
dataInconsistency Skip if source/sink file changed by other concurrent write. Default is false. Type: boolean (or Expression with resultType boolean).
fileMissing Skip if file is deleted by other client during copy. Default is true. Type: boolean (or Expression with resultType boolean).

CopySource

Name Description Value
disableMetricsCollection If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean).
maxConcurrentConnections The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer).
sourceRetryCount Source retry count. Type: integer (or Expression with resultType integer).
sourceRetryWait Source retry wait. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
type Set the object type AmazonMWSSource
AmazonRdsForOracleSource
AmazonRdsForSqlServerSource
AmazonRedshiftSource
AvroSource
AzureBlobFSSource
AzureDatabricksDeltaLakeSource
AzureDataExplorerSource
AzureDataLakeStoreSource
AzureMariaDBSource
AzureMySqlSource
AzurePostgreSqlSource
AzureSqlSource
AzureTableSource
BinarySource
BlobSource
CassandraSource
CommonDataServiceForAppsSource
ConcurSource
CosmosDbMongoDbApiSource
CosmosDbSqlApiSource
CouchbaseSource
Db2Source
DelimitedTextSource
DocumentDbCollectionSource
DrillSource
DynamicsAXSource
DynamicsCrmSource
DynamicsSource
EloquaSource
ExcelSource
FileSystemSource
GoogleAdWordsSource
GoogleBigQuerySource
GoogleBigQueryV2Source
GreenplumSource
HBaseSource
HdfsSource
HiveSource
HttpSource
HubspotSource
ImpalaSource
InformixSource
JiraSource
JsonSource
LakeHouseTableSource
MagentoSource
MariaDBSource
MarketoSource
MicrosoftAccessSource
MongoDbAtlasSource
MongoDbSource
MongoDbV2Source
MySqlSource
NetezzaSource
ODataSource
OdbcSource
Office365Source
OracleServiceCloudSource
OracleSource
OrcSource
ParquetSource
PaypalSource
PhoenixSource
PostgreSqlSource
PostgreSqlV2Source
PrestoSource
QuickBooksSource
RelationalSource
ResponsysSource
RestSource
SalesforceMarketingCloudSource
SalesforceServiceCloudSource
SalesforceServiceCloudV2Source
SalesforceSource
SalesforceV2Source
SapBwSource
SapCloudForCustomerSource
SapEccSource
SapHanaSource
SapOdpSource
SapOpenHubSource
SapTableSource
ServiceNowSource
ServiceNowV2Source
SharePointOnlineListSource
ShopifySource
SnowflakeSource
SnowflakeV2Source
SparkSource
SqlDWSource
SqlMISource
SqlServerSource
SqlSource
SquareSource
SybaseSource
TeradataSource
VerticaSource
WarehouseSource
WebSource
XeroSource
XmlSource
ZohoSource (required)

AmazonMWSSource

Name Description Value
type Copy source type. 'AmazonMWSSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

AmazonRdsForOracleSource

Name Description Value
type Copy source type. 'AmazonRdsForOracleSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
oracleReaderQuery AmazonRdsForOracle reader query. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for AmazonRdsForOracle read in parallel. Type: string (or Expression with resultType string).
partitionSettings The settings that will be leveraged for AmazonRdsForOracle source partitioning. AmazonRdsForOraclePartitionSettings
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

AmazonRdsForOraclePartitionSettings

Name Description Value
partitionColumnName The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).
partitionLowerBound The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).
partitionNames Names of the physical partitions of AmazonRdsForOracle table.
partitionUpperBound The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).

AmazonRdsForSqlServerSource

Name Description Value
type Copy source type. 'AmazonRdsForSqlServerSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
isolationLevel Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange".
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
produceAdditionalTypes Which additional types to produce.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
sqlReaderQuery SQL reader query. Type: string (or Expression with resultType string).
sqlReaderStoredProcedureName Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string).
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}".

SqlPartitionSettings

Name Description Value
partitionColumnName The name of the column in integer or datetime type that will be used for proceeding partitioning. If not specified, the primary key of the table is auto-detected and used as the partition column. Type: string (or Expression with resultType string).
partitionLowerBound The minimum value of the partition column for partition range splitting. This value is used to decide the partition stride, not for filtering the rows in table. All rows in the table or query result will be partitioned and copied. Type: string (or Expression with resultType string).
partitionUpperBound The maximum value of the partition column for partition range splitting. This value is used to decide the partition stride, not for filtering the rows in table. All rows in the table or query result will be partitioned and copied. Type: string (or Expression with resultType string).

AmazonRedshiftSource

Name Description Value
type Copy source type. 'AmazonRedshiftSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
redshiftUnloadSettings The Amazon S3 settings needed for the interim Amazon S3 when copying from Amazon Redshift with unload. With this, data from Amazon Redshift source will be unloaded into S3 first and then copied into the targeted sink from the interim S3. RedshiftUnloadSettings

RedshiftUnloadSettings

Name Description Value
bucketName The bucket of the interim Amazon S3 which will be used to store the unloaded data from Amazon Redshift source. The bucket must be in the same region as the Amazon Redshift source. Type: string (or Expression with resultType string).
s3LinkedServiceName The name of the Amazon S3 linked service which will be used for the unload operation when copying from the Amazon Redshift source. LinkedServiceReference (required)

AvroSource

Name Description Value
type Copy source type. 'AvroSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
storeSettings Avro store settings. StoreReadSettings

StoreReadSettings

Name Description Value
disableMetricsCollection If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean).
maxConcurrentConnections The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer).
type Set the object type AmazonS3CompatibleReadSettings
AmazonS3ReadSettings
AzureBlobFSReadSettings
AzureBlobStorageReadSettings
AzureDataLakeStoreReadSettings
AzureFileStorageReadSettings
FileServerReadSettings
FtpReadSettings
GoogleCloudStorageReadSettings
HdfsReadSettings
HttpReadSettings
LakeHouseReadSettings
OracleCloudStorageReadSettings
SftpReadSettings (required)

AmazonS3CompatibleReadSettings

Name Description Value
type The read setting type. 'AmazonS3CompatibleReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
prefix The prefix filter for the S3 Compatible object name. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName Amazon S3 Compatible wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath Amazon S3 Compatible wildcardFolderPath. Type: string (or Expression with resultType string).

AmazonS3ReadSettings

Name Description Value
type The read setting type. 'AmazonS3ReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
prefix The prefix filter for the S3 object name. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName AmazonS3 wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath AmazonS3 wildcardFolderPath. Type: string (or Expression with resultType string).

AzureBlobFSReadSettings

Name Description Value
type The read setting type. 'AzureBlobFSReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName Azure blobFS wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath Azure blobFS wildcardFolderPath. Type: string (or Expression with resultType string).

AzureBlobStorageReadSettings

Name Description Value
type The read setting type. 'AzureBlobStorageReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
prefix The prefix filter for the Azure Blob name. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName Azure blob wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath Azure blob wildcardFolderPath. Type: string (or Expression with resultType string).

AzureDataLakeStoreReadSettings

Name Description Value
type The read setting type. 'AzureDataLakeStoreReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
listAfter Lists files after the value (exclusive) based on file/folder names’ lexicographical order. Applies under the folderPath in data set, and filter files/sub-folders under the folderPath. Type: string (or Expression with resultType string).
listBefore Lists files before the value (inclusive) based on file/folder names’ lexicographical order. Applies under the folderPath in data set, and filter files/sub-folders under the folderPath. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName ADLS wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath ADLS wildcardFolderPath. Type: string (or Expression with resultType string).

AzureFileStorageReadSettings

Name Description Value
type The read setting type. 'AzureFileStorageReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
prefix The prefix filter for the Azure File name starting from root path. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName Azure File Storage wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath Azure File Storage wildcardFolderPath. Type: string (or Expression with resultType string).

FileServerReadSettings

Name Description Value
type The read setting type. 'FileServerReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileFilter Specify a filter to be used to select a subset of files in the folderPath rather than all files. Type: string (or Expression with resultType string).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName FileServer wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath FileServer wildcardFolderPath. Type: string (or Expression with resultType string).

FtpReadSettings

Name Description Value
type The read setting type. 'FtpReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
disableChunking If true, disable parallel reading within each file. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
useBinaryTransfer Specify whether to use binary transfer mode for FTP stores. Type: boolean (or Expression with resultType boolean).
wildcardFileName Ftp wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath Ftp wildcardFolderPath. Type: string (or Expression with resultType string).

GoogleCloudStorageReadSettings

Name Description Value
type The read setting type. 'GoogleCloudStorageReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
prefix The prefix filter for the Google Cloud Storage object name. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName Google Cloud Storage wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath Google Cloud Storage wildcardFolderPath. Type: string (or Expression with resultType string).

HdfsReadSettings

Name Description Value
type The read setting type. 'HdfsReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
distcpSettings Specifies Distcp-related settings. DistcpSettings
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName HDFS wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath HDFS wildcardFolderPath. Type: string (or Expression with resultType string).

DistcpSettings

Name Description Value
distcpOptions Specifies the Distcp options. Type: string (or Expression with resultType string).
resourceManagerEndpoint Specifies the Yarn ResourceManager endpoint. Type: string (or Expression with resultType string).
tempScriptPath Specifies an existing folder path which will be used to store temp Distcp command script. The script file is generated by ADF and will be removed after Copy job finished. Type: string (or Expression with resultType string).

HttpReadSettings

Name Description Value
type The read setting type. 'HttpReadSettings' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
additionalHeaders The additional HTTP headers in the request to the RESTful API. Type: string (or Expression with resultType string).
requestBody The HTTP request body to the RESTful API if requestMethod is POST. Type: string (or Expression with resultType string).
requestMethod The HTTP method used to call the RESTful API. The default is GET. Type: string (or Expression with resultType string).
requestTimeout Specifies the timeout for a HTTP client to get HTTP response from HTTP server. Type: string (or Expression with resultType string).

LakeHouseReadSettings

Name Description Value
type The read setting type. 'LakeHouseReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName Microsoft Fabric LakeHouse Files wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath Microsoft Fabric LakeHouse Files wildcardFolderPath. Type: string (or Expression with resultType string).

OracleCloudStorageReadSettings

Name Description Value
type The read setting type. 'OracleCloudStorageReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
prefix The prefix filter for the Oracle Cloud Storage object name. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName Oracle Cloud Storage wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath Oracle Cloud Storage wildcardFolderPath. Type: string (or Expression with resultType string).

SftpReadSettings

Name Description Value
type The read setting type. 'SftpReadSettings' (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
disableChunking If true, disable parallel reading within each file. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName Sftp wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath Sftp wildcardFolderPath. Type: string (or Expression with resultType string).

AzureBlobFSSource

Name Description Value
type Copy source type. 'AzureBlobFSSource' (required)
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
skipHeaderLineCount Number of header lines to skip from each blob. Type: integer (or Expression with resultType integer).
treatEmptyAsNull Treat empty as null. Type: boolean (or Expression with resultType boolean).

AzureDatabricksDeltaLakeSource

Name Description Value
type Copy source type. 'AzureDatabricksDeltaLakeSource' (required)
exportSettings Azure Databricks Delta Lake export settings. AzureDatabricksDeltaLakeExportCommand
query Azure Databricks Delta Lake Sql query. Type: string (or Expression with resultType string).

AzureDatabricksDeltaLakeExportCommand

Name Description Value
dateFormat Specify the date format for the csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string).
timestampFormat Specify the timestamp format for the csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string).
type The export setting type. string (required)
{customized property}

AzureDataExplorerSource

Name Description Value
type Copy source type. 'AzureDataExplorerSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
noTruncation The name of the Boolean option that controls whether truncation is applied to result-sets that go beyond a certain row-count limit.
query Database query. Should be a Kusto Query Language (KQL) query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9]))..

AzureDataLakeStoreSource

Name Description Value
type Copy source type. 'AzureDataLakeStoreSource' (required)
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).

AzureMariaDBSource

Name Description Value
type Copy source type. 'AzureMariaDBSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

AzureMySqlSource

Name Description Value
type Copy source type. 'AzureMySqlSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

AzurePostgreSqlSource

Name Description Value
type Copy source type. 'AzurePostgreSqlSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

AzureSqlSource

Name Description Value
type Copy source type. 'AzureSqlSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
isolationLevel Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string).
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
produceAdditionalTypes Which additional types to produce.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
sqlReaderQuery SQL reader query. Type: string (or Expression with resultType string).
sqlReaderStoredProcedureName Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string).
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}".

AzureTableSource

Name Description Value
type Copy source type. 'AzureTableSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
azureTableSourceIgnoreTableNotFound Azure Table source ignore table not found. Type: boolean (or Expression with resultType boolean).
azureTableSourceQuery Azure Table source query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

BinarySource

Name Description Value
type Copy source type. 'BinarySource' (required)
formatSettings Binary format settings. BinaryReadSettings
storeSettings Binary store settings. StoreReadSettings

BinaryReadSettings

Name Description Value
compressionProperties Compression settings. CompressionReadSettings
type The read setting type. string (required)
{customized property}

CompressionReadSettings

Name Description Value
type Set the object type TarGZipReadSettings
TarReadSettings
ZipDeflateReadSettings (required)

TarGZipReadSettings

Name Description Value
type The Compression setting type. 'TarGZipReadSettings' (required)
preserveCompressionFileNameAsFolder Preserve the compression file name as folder path. Type: boolean (or Expression with resultType boolean).

TarReadSettings

Name Description Value
type The Compression setting type. 'TarReadSettings' (required)
preserveCompressionFileNameAsFolder Preserve the compression file name as folder path. Type: boolean (or Expression with resultType boolean).

ZipDeflateReadSettings

Name Description Value
type The Compression setting type. 'ZipDeflateReadSettings' (required)
preserveZipFileNameAsFolder Preserve the zip file name as folder path. Type: boolean (or Expression with resultType boolean).

BlobSource

Name Description Value
type Copy source type. 'BlobSource' (required)
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
skipHeaderLineCount Number of header lines to skip from each blob. Type: integer (or Expression with resultType integer).
treatEmptyAsNull Treat empty as null. Type: boolean (or Expression with resultType boolean).

CassandraSource

Name Description Value
type Copy source type. 'CassandraSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
consistencyLevel The consistency level specifies how many Cassandra servers must respond to a read request before returning data to the client application. Cassandra checks the specified number of Cassandra servers for data to satisfy the read request. Must be one of cassandraSourceReadConsistencyLevels. The default value is 'ONE'. It is case-insensitive. 'ALL'
'EACH_QUORUM'
'LOCAL_ONE'
'LOCAL_QUORUM'
'LOCAL_SERIAL'
'ONE'
'QUORUM'
'SERIAL'
'THREE'
'TWO'
query Database query. Should be a SQL-92 query expression or Cassandra Query Language (CQL) command. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

CommonDataServiceForAppsSource

Name Description Value
type Copy source type. 'CommonDataServiceForAppsSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query FetchXML is a proprietary query language that is used in Microsoft Common Data Service for Apps (online & on-premises). Type: string (or Expression with resultType string).

ConcurSource

Name Description Value
type Copy source type. 'ConcurSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

CosmosDbMongoDbApiSource

Name Description Value
type Copy source type. 'CosmosDbMongoDbApiSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
batchSize Specifies the number of documents to return in each batch of the response from MongoDB instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer).
cursorMethods Cursor methods for Mongodb query. MongoDbCursorMethodsProperties
filter Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

MongoDbCursorMethodsProperties

Name Description Value
limit Specifies the maximum number of documents the server returns. limit() is analogous to the LIMIT statement in a SQL database. Type: integer (or Expression with resultType integer).
project Specifies the fields to return in the documents that match the query filter. To return all fields in the matching documents, omit this parameter. Type: string (or Expression with resultType string).
skip Specifies the how many documents skipped and where MongoDB begins returning results. This approach may be useful in implementing paginated results. Type: integer (or Expression with resultType integer).
sort Specifies the order in which the query returns matching documents. Type: string (or Expression with resultType string). Type: string (or Expression with resultType string).
{customized property}

CosmosDbSqlApiSource

Name Description Value
type Copy source type. 'CosmosDbSqlApiSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
detectDatetime Whether detect primitive values as datetime values. Type: boolean (or Expression with resultType boolean).
pageSize Page size of the result. Type: integer (or Expression with resultType integer).
preferredRegions Preferred regions. Type: array of strings (or Expression with resultType array of strings).
query SQL API query. Type: string (or Expression with resultType string).

CouchbaseSource

Name Description Value
type Copy source type. 'CouchbaseSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

Db2Source

Name Description Value
type Copy source type. 'Db2Source' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

DelimitedTextSource

Name Description Value
type Copy source type. 'DelimitedTextSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
formatSettings DelimitedText format settings. DelimitedTextReadSettings
storeSettings DelimitedText store settings. StoreReadSettings

DelimitedTextReadSettings

Name Description Value
compressionProperties Compression settings. CompressionReadSettings
skipLineCount Indicates the number of non-empty rows to skip when reading data from input files. Type: integer (or Expression with resultType integer).
type The read setting type. string (required)
{customized property}

DocumentDbCollectionSource

Name Description Value
type Copy source type. 'DocumentDbCollectionSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
nestingSeparator Nested properties separator. Type: string (or Expression with resultType string).
query Documents query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

DrillSource

Name Description Value
type Copy source type. 'DrillSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

DynamicsAXSource

Name Description Value
type Copy source type. 'DynamicsAXSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

DynamicsCrmSource

Name Description Value
type Copy source type. 'DynamicsCrmSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query FetchXML is a proprietary query language that is used in Microsoft Dynamics CRM (online & on-premises). Type: string (or Expression with resultType string).

DynamicsSource

Name Description Value
type Copy source type. 'DynamicsSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query FetchXML is a proprietary query language that is used in Microsoft Dynamics (online & on-premises). Type: string (or Expression with resultType string).

EloquaSource

Name Description Value
type Copy source type. 'EloquaSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

ExcelSource

Name Description Value
type Copy source type. 'ExcelSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
storeSettings Excel store settings. StoreReadSettings

FileSystemSource

Name Description Value
type Copy source type. 'FileSystemSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).

GoogleAdWordsSource

Name Description Value
type Copy source type. 'GoogleAdWordsSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

GoogleBigQuerySource

Name Description Value
type Copy source type. 'GoogleBigQuerySource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

GoogleBigQueryV2Source

Name Description Value
type Copy source type. 'GoogleBigQueryV2Source' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

GreenplumSource

Name Description Value
type Copy source type. 'GreenplumSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

HBaseSource

Name Description Value
type Copy source type. 'HBaseSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

HdfsSource

Name Description Value
type Copy source type. 'HdfsSource' (required)
distcpSettings Specifies Distcp-related settings. DistcpSettings
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).

HiveSource

Name Description Value
type Copy source type. 'HiveSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

HttpSource

Name Description Value
type Copy source type. 'HttpSource' (required)
httpRequestTimeout Specifies the timeout for a HTTP client to get HTTP response from HTTP server. The default value is equivalent to System.Net.HttpWebRequest.Timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

HubspotSource

Name Description Value
type Copy source type. 'HubspotSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

ImpalaSource

Name Description Value
type Copy source type. 'ImpalaSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

InformixSource

Name Description Value
type Copy source type. 'InformixSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

JiraSource

Name Description Value
type Copy source type. 'JiraSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

JsonSource

Name Description Value
type Copy source type. 'JsonSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
formatSettings Json format settings. JsonReadSettings
storeSettings Json store settings. StoreReadSettings

JsonReadSettings

Name Description Value
compressionProperties Compression settings. CompressionReadSettings
type The read setting type. string (required)
{customized property}

LakeHouseTableSource

Name Description Value
type Copy source type. 'LakeHouseTableSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
timestampAsOf Query an older snapshot by timestamp. Type: string (or Expression with resultType string).
versionAsOf Query an older snapshot by version. Type: integer (or Expression with resultType integer).

MagentoSource

Name Description Value
type Copy source type. 'MagentoSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

MariaDBSource

Name Description Value
type Copy source type. 'MariaDBSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

MarketoSource

Name Description Value
type Copy source type. 'MarketoSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

MicrosoftAccessSource

Name Description Value
type Copy source type. 'MicrosoftAccessSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).

MongoDbAtlasSource

Name Description Value
type Copy source type. 'MongoDbAtlasSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
batchSize Specifies the number of documents to return in each batch of the response from MongoDB Atlas instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer).
cursorMethods Cursor methods for Mongodb query MongoDbCursorMethodsProperties
filter Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

MongoDbSource

Name Description Value
type Copy source type. 'MongoDbSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Should be a SQL-92 query expression. Type: string (or Expression with resultType string).

MongoDbV2Source

Name Description Value
type Copy source type. 'MongoDbV2Source' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
batchSize Specifies the number of documents to return in each batch of the response from MongoDB instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer).
cursorMethods Cursor methods for Mongodb query MongoDbCursorMethodsProperties
filter Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

MySqlSource

Name Description Value
type Copy source type. 'MySqlSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

NetezzaSource

Name Description Value
type Copy source type. 'NetezzaSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
partitionOption The partition mechanism that will be used for Netezza read in parallel. Possible values include: "None", "DataSlice", "DynamicRange".
partitionSettings The settings that will be leveraged for Netezza source partitioning. NetezzaPartitionSettings
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

NetezzaPartitionSettings

Name Description Value
partitionColumnName The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).
partitionLowerBound The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).
partitionUpperBound The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).

ODataSource

Name Description Value
type Copy source type. 'ODataSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
query OData query. For example, "$top=1". Type: string (or Expression with resultType string).

OdbcSource

Name Description Value
type Copy source type. 'OdbcSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

Office365Source

Name Description Value
type Copy source type. 'Office365Source' (required)
allowedGroups The groups containing all the users. Type: array of strings (or Expression with resultType array of strings).
dateFilterColumn The Column to apply the {paramref name="StartTime"/} and {paramref name="EndTime"/}. Type: string (or Expression with resultType string).
endTime End time of the requested range for this dataset. Type: string (or Expression with resultType string).
outputColumns The columns to be read out from the Office 365 table. Type: array of objects (or Expression with resultType array of objects). itemType: OutputColumn. Example: [ { "name": "Id" }, { "name": "CreatedDateTime" } ]
startTime Start time of the requested range for this dataset. Type: string (or Expression with resultType string).
userScopeFilterUri The user scope uri. Type: string (or Expression with resultType string).

OracleServiceCloudSource

Name Description Value
type Copy source type. 'OracleServiceCloudSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

OracleSource

Name Description Value
type Copy source type. 'OracleSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
oracleReaderQuery Oracle reader query. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for Oracle read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange".
partitionSettings The settings that will be leveraged for Oracle source partitioning. OraclePartitionSettings
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

OraclePartitionSettings

Name Description Value
partitionColumnName The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).
partitionLowerBound The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).
partitionNames Names of the physical partitions of Oracle table.
partitionUpperBound The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).

OrcSource

Name Description Value
type Copy source type. 'OrcSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
storeSettings ORC store settings. StoreReadSettings

ParquetSource

Name Description Value
type Copy source type. 'ParquetSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
formatSettings Parquet format settings. ParquetReadSettings
storeSettings Parquet store settings. StoreReadSettings

ParquetReadSettings

Name Description Value
compressionProperties Compression settings. CompressionReadSettings
type The read setting type. string (required)
{customized property}

PaypalSource

Name Description Value
type Copy source type. 'PaypalSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

PhoenixSource

Name Description Value
type Copy source type. 'PhoenixSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

PostgreSqlSource

Name Description Value
type Copy source type. 'PostgreSqlSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

PostgreSqlV2Source

Name Description Value
type Copy source type. 'PostgreSqlV2Source' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

PrestoSource

Name Description Value
type Copy source type. 'PrestoSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

QuickBooksSource

Name Description Value
type Copy source type. 'QuickBooksSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

RelationalSource

Name Description Value
type Copy source type. 'RelationalSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).

ResponsysSource

Name Description Value
type Copy source type. 'ResponsysSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

RestSource

Name Description Value
type Copy source type. 'RestSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: key value pairs (value should be string type).
additionalHeaders The additional HTTP headers in the request to the RESTful API. Type: string (or Expression with resultType string).
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
paginationRules The pagination rules to compose next page requests. Type: string (or Expression with resultType string).
requestBody The HTTP request body to the RESTful API if requestMethod is POST. Type: string (or Expression with resultType string).
requestInterval The time to await before sending next page request.
requestMethod The HTTP method used to call the RESTful API. The default is GET. Type: string (or Expression with resultType string).

SalesforceMarketingCloudSource

Name Description Value
type Copy source type. 'SalesforceMarketingCloudSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

SalesforceServiceCloudSource

Name Description Value
type Copy source type. 'SalesforceServiceCloudSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
readBehavior The read behavior for the operation. Default is Query. Allowed values: Query/QueryAll. Type: string (or Expression with resultType string).

SalesforceServiceCloudV2Source

Name Description Value
type Copy source type. 'SalesforceServiceCloudV2Source' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
includeDeletedObjects This property control whether query result contains Deleted objects. Default is false. Type: boolean (or Expression with resultType boolean).
SOQLQuery Database query. Type: string (or Expression with resultType string).

SalesforceSource

Name Description Value
type Copy source type. 'SalesforceSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
readBehavior The read behavior for the operation. Default is Query. Allowed values: Query/QueryAll. Type: string (or Expression with resultType string).

SalesforceV2Source

Name Description Value
type Copy source type. 'SalesforceV2Source' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
includeDeletedObjects This property control whether query result contains Deleted objects. Default is false. Type: boolean (or Expression with resultType boolean).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
SOQLQuery Database query. Type: string (or Expression with resultType string).

SapBwSource

Name Description Value
type Copy source type. 'SapBwSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query MDX query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

SapCloudForCustomerSource

Name Description Value
type Copy source type. 'SapCloudForCustomerSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
query SAP Cloud for Customer OData query. For example, "$top=1". Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

SapEccSource

Name Description Value
type Copy source type. 'SapEccSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
query SAP ECC OData query. For example, "$top=1". Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

SapHanaSource

Name Description Value
type Copy source type. 'SapHanaSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
packetSize The packet size of data read from SAP HANA. Type: integer(or Expression with resultType integer).
partitionOption The partition mechanism that will be used for SAP HANA read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "SapHanaDynamicRange".
partitionSettings The settings that will be leveraged for SAP HANA source partitioning. SapHanaPartitionSettings
query SAP HANA Sql query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

SapHanaPartitionSettings

Name Description Value
partitionColumnName The name of the column that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).

SapOdpSource

Name Description Value
type Copy source type. 'SapOdpSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
extractionMode The extraction mode. Allowed value include: Full, Delta and Recovery. The default value is Full. Type: string (or Expression with resultType string).
projection Specifies the columns to be selected from source data. Type: array of objects(projection) (or Expression with resultType array of objects).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
selection Specifies the selection conditions from source data. Type: array of objects(selection) (or Expression with resultType array of objects).
subscriberProcess The subscriber process to manage the delta process. Type: string (or Expression with resultType string).

SapOpenHubSource

Name Description Value
type Copy source type. 'SapOpenHubSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
baseRequestId The ID of request for delta loading. Once it is set, only data with requestId larger than the value of this property will be retrieved. The default value is 0. Type: integer (or Expression with resultType integer ).
customRfcReadTableFunctionModule Specifies the custom RFC function module that will be used to read data from SAP Table. Type: string (or Expression with resultType string).
excludeLastRequest Whether to exclude the records of the last request. The default value is true. Type: boolean (or Expression with resultType boolean).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
sapDataColumnDelimiter The single character that will be used as delimiter passed to SAP RFC as well as splitting the output data retrieved. Type: string (or Expression with resultType string).

SapTableSource

Name Description Value
type Copy source type. 'SapTableSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
batchSize Specifies the maximum number of rows that will be retrieved at a time when retrieving data from SAP Table. Type: integer (or Expression with resultType integer).
customRfcReadTableFunctionModule Specifies the custom RFC function module that will be used to read data from SAP Table. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for SAP table read in parallel. Possible values include: "None", "PartitionOnInt", "PartitionOnCalendarYear", "PartitionOnCalendarMonth", "PartitionOnCalendarDate", "PartitionOnTime".
partitionSettings The settings that will be leveraged for SAP table source partitioning. SapTablePartitionSettings
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
rfcTableFields The fields of the SAP table that will be retrieved. For example, column0, column1. Type: string (or Expression with resultType string).
rfcTableOptions The options for the filtering of the SAP Table. For example, COLUMN0 EQ SOME VALUE. Type: string (or Expression with resultType string).
rowCount The number of rows to be retrieved. Type: integer(or Expression with resultType integer).
rowSkips The number of rows that will be skipped. Type: integer (or Expression with resultType integer).
sapDataColumnDelimiter The single character that will be used as delimiter passed to SAP RFC as well as splitting the output data retrieved. Type: string (or Expression with resultType string).

SapTablePartitionSettings

Name Description Value
maxPartitionsNumber The maximum value of partitions the table will be split into. Type: integer (or Expression with resultType string).
partitionColumnName The name of the column that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).
partitionLowerBound The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).
partitionUpperBound The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).

ServiceNowSource

Name Description Value
type Copy source type. 'ServiceNowSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

ServiceNowV2Source

Name Description Value
type Copy source type. 'ServiceNowV2Source' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
expression Expression to filter data from source. ExpressionV2
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

ExpressionV2

Name Description Value
operands List of nested expressions. ExpressionV2[]
operator Expression operator value Type: string. string
type Type of expressions supported by the system. Type: string. 'Binary'
'Constant'
'Field'
'Unary'
value Value for Constant/Field Type: string. string

SharePointOnlineListSource

Name Description Value
type Copy source type. 'SharePointOnlineListSource' (required)
httpRequestTimeout The wait time to get a response from SharePoint Online. Default value is 5 minutes (00:05:00). Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
query The OData query to filter the data in SharePoint Online list. For example, "$top=1". Type: string (or Expression with resultType string).

ShopifySource

Name Description Value
type Copy source type. 'ShopifySource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

SnowflakeSource

Name Description Value
type Copy source type. 'SnowflakeSource' (required)
exportSettings Snowflake export settings. SnowflakeExportCopyCommand (required)
query Snowflake Sql query. Type: string (or Expression with resultType string).

SnowflakeExportCopyCommand

Name Description Value
additionalCopyOptions Additional copy options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": "'HH24:MI:SS.FF'" } SnowflakeExportCopyCommandAdditionalCopyOptions
additionalFormatOptions Additional format options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalFormatOptions": { "OVERWRITE": "TRUE", "MAX_FILE_SIZE": "'FALSE'" } SnowflakeExportCopyCommandAdditionalFormatOptions
type The export setting type. string (required)
{customized property}

SnowflakeExportCopyCommandAdditionalCopyOptions

Name Description Value
{customized property}

SnowflakeExportCopyCommandAdditionalFormatOptions

Name Description Value
{customized property}

SnowflakeV2Source

Name Description Value
type Copy source type. 'SnowflakeV2Source' (required)
exportSettings Snowflake export settings. SnowflakeExportCopyCommand (required)
query Snowflake Sql query. Type: string (or Expression with resultType string).

SparkSource

Name Description Value
type Copy source type. 'SparkSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

SqlDWSource

Name Description Value
type Copy source type. 'SqlDWSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
isolationLevel Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string).
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
sqlReaderQuery SQL Data Warehouse reader query. Type: string (or Expression with resultType string).
sqlReaderStoredProcedureName Name of the stored procedure for a SQL Data Warehouse source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string).
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". Type: object (or Expression with resultType object), itemType: StoredProcedureParameter.

SqlMISource

Name Description Value
type Copy source type. 'SqlMISource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
isolationLevel Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string).
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
produceAdditionalTypes Which additional types to produce.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
sqlReaderQuery SQL reader query. Type: string (or Expression with resultType string).
sqlReaderStoredProcedureName Name of the stored procedure for a Azure SQL Managed Instance source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string).
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}".

SqlServerSource

Name Description Value
type Copy source type. 'SqlServerSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
isolationLevel Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string).
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
produceAdditionalTypes Which additional types to produce.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
sqlReaderQuery SQL reader query. Type: string (or Expression with resultType string).
sqlReaderStoredProcedureName Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string).
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}".

SqlSource

Name Description Value
type Copy source type. 'SqlSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
isolationLevel Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string).
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
sqlReaderQuery SQL reader query. Type: string (or Expression with resultType string).
sqlReaderStoredProcedureName Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string).
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}".

SquareSource

Name Description Value
type Copy source type. 'SquareSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

SybaseSource

Name Description Value
type Copy source type. 'SybaseSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

TeradataSource

Name Description Value
type Copy source type. 'TeradataSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
partitionOption The partition mechanism that will be used for teradata read in parallel. Possible values include: "None", "Hash", "DynamicRange".
partitionSettings The settings that will be leveraged for teradata source partitioning. TeradataPartitionSettings
query Teradata query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

TeradataPartitionSettings

Name Description Value
partitionColumnName The name of the column that will be used for proceeding range or hash partitioning. Type: string (or Expression with resultType string).
partitionLowerBound The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).
partitionUpperBound The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).

VerticaSource

Name Description Value
type Copy source type. 'VerticaSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

WarehouseSource

Name Description Value
type Copy source type. 'WarehouseSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
isolationLevel Specifies the transaction locking behavior for the Microsoft Fabric Warehouse source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange".
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
sqlReaderQuery Microsoft Fabric Warehouse reader query. Type: string (or Expression with resultType string).
sqlReaderStoredProcedureName Name of the stored procedure for a Microsoft Fabric Warehouse source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string).
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". Type: object (or Expression with resultType object), itemType: StoredProcedureParameter.

WebSource

Name Description Value
type Copy source type. 'WebSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).

XeroSource

Name Description Value
type Copy source type. 'XeroSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

XmlSource

Name Description Value
type Copy source type. 'XmlSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
formatSettings Xml format settings. XmlReadSettings
storeSettings Xml store settings. StoreReadSettings

XmlReadSettings

Name Description Value
compressionProperties Compression settings. CompressionReadSettings
detectDataType Indicates whether type detection is enabled when reading the xml files. Type: boolean (or Expression with resultType boolean).
namespacePrefixes Namespace uri to prefix mappings to override the prefixes in column names when namespace is enabled, if no prefix is defined for a namespace uri, the prefix of xml element/attribute name in the xml data file will be used. Example: "{"http://www.example.com/xml":"prefix"}" Type: object (or Expression with resultType object).
namespaces Indicates whether namespace is enabled when reading the xml files. Type: boolean (or Expression with resultType boolean).
type The read setting type. string (required)
validationMode Indicates what validation method is used when reading the xml files. Allowed values: 'none', 'xsd', or 'dtd'. Type: string (or Expression with resultType string).
{customized property}

ZohoSource

Name Description Value
type Copy source type. 'ZohoSource' (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

StagingSettings

Name Description Value
enableCompression Specifies whether to use compression when copying data via an interim staging. Default value is false. Type: boolean (or Expression with resultType boolean).
linkedServiceName Staging linked service reference. LinkedServiceReference (required)
path The path to storage for storing the interim data. Type: string (or Expression with resultType string).
{customized property}

CustomActivity

Name Description Value
type Type of activity. 'Custom' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Custom activity properties. CustomActivityTypeProperties (required)

CustomActivityTypeProperties

Name Description Value
autoUserSpecification Elevation level and scope for the user, default is nonadmin task. Type: string (or Expression with resultType double).
command Command for custom activity Type: string (or Expression with resultType string).
extendedProperties User defined property bag. There is no restriction on the keys or values that can be used. The user specified custom activity has the full responsibility to consume and interpret the content defined. CustomActivityTypePropertiesExtendedProperties
folderPath Folder path for resource files Type: string (or Expression with resultType string).
referenceObjects Reference objects CustomActivityReferenceObject
resourceLinkedService Resource linked service reference. LinkedServiceReference
retentionTimeInDays The retention time for the files submitted for custom activity. Type: double (or Expression with resultType double).

CustomActivityTypePropertiesExtendedProperties

Name Description Value
{customized property}

CustomActivityReferenceObject

Name Description Value
datasets Dataset references. DatasetReference[]
linkedServices Linked service references. LinkedServiceReference[]

DatabricksNotebookActivity

Name Description Value
type Type of activity. 'DatabricksNotebook' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Databricks Notebook activity properties. DatabricksNotebookActivityTypeProperties (required)

DatabricksNotebookActivityTypeProperties

Name Description Value
baseParameters Base parameters to be used for each run of this job.If the notebook takes a parameter that is not specified, the default value from the notebook will be used. DatabricksNotebookActivityTypePropertiesBaseParamete...
libraries A list of libraries to be installed on the cluster that will execute the job. DatabricksNotebookActivityTypePropertiesLibrariesIte...[]
notebookPath The absolute path of the notebook to be run in the Databricks Workspace. This path must begin with a slash. Type: string (or Expression with resultType string).

DatabricksNotebookActivityTypePropertiesBaseParamete...

Name Description Value
{customized property}

DatabricksNotebookActivityTypePropertiesLibrariesIte...

Name Description Value
{customized property}

DatabricksSparkJarActivity

Name Description Value
type Type of activity. 'DatabricksSparkJar' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Databricks SparkJar activity properties. DatabricksSparkJarActivityTypeProperties (required)

DatabricksSparkJarActivityTypeProperties

Name Description Value
libraries A list of libraries to be installed on the cluster that will execute the job. DatabricksSparkJarActivityTypePropertiesLibrariesIte...[]
mainClassName The full name of the class containing the main method to be executed. This class must be contained in a JAR provided as a library. Type: string (or Expression with resultType string).
parameters Parameters that will be passed to the main method. any[]

DatabricksSparkJarActivityTypePropertiesLibrariesIte...

Name Description Value
{customized property}

DatabricksSparkPythonActivity

Name Description Value
type Type of activity. 'DatabricksSparkPython' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Databricks SparkPython activity properties. DatabricksSparkPythonActivityTypeProperties (required)

DatabricksSparkPythonActivityTypeProperties

Name Description Value
libraries A list of libraries to be installed on the cluster that will execute the job. DatabricksSparkPythonActivityTypePropertiesLibraries...[]
parameters Command line parameters that will be passed to the Python file. any[]
pythonFile The URI of the Python file to be executed. DBFS paths are supported. Type: string (or Expression with resultType string).

DatabricksSparkPythonActivityTypePropertiesLibraries...

Name Description Value
{customized property}

DataLakeAnalyticsUsqlActivity

Name Description Value
type Type of activity. 'DataLakeAnalyticsU-SQL' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Data Lake Analytics U-SQL activity properties. DataLakeAnalyticsUsqlActivityTypeProperties (required)

DataLakeAnalyticsUsqlActivityTypeProperties

Name Description Value
compilationMode Compilation mode of U-SQL. Must be one of these values : Semantic, Full and SingleBox. Type: string (or Expression with resultType string).
degreeOfParallelism The maximum number of nodes simultaneously used to run the job. Default value is 1. Type: integer (or Expression with resultType integer), minimum: 1.
parameters Parameters for U-SQL job request. DataLakeAnalyticsUsqlActivityTypePropertiesParameter...
priority Determines which jobs out of all that are queued should be selected to run first. The lower the number, the higher the priority. Default value is 1000. Type: integer (or Expression with resultType integer), minimum: 1.
runtimeVersion Runtime version of the U-SQL engine to use. Type: string (or Expression with resultType string).
scriptLinkedService Script linked service reference. LinkedServiceReference (required)
scriptPath Case-sensitive path to folder that contains the U-SQL script. Type: string (or Expression with resultType string).

DataLakeAnalyticsUsqlActivityTypePropertiesParameter...

Name Description Value
{customized property}

DeleteActivity

Name Description Value
type Type of activity. 'Delete' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Delete activity properties. DeleteActivityTypeProperties (required)

DeleteActivityTypeProperties

Name Description Value
dataset Delete activity dataset reference. DatasetReference (required)
enableLogging Whether to record detailed logs of delete-activity execution. Default value is false. Type: boolean (or Expression with resultType boolean).
logStorageSettings Log storage settings customer need to provide when enableLogging is true. LogStorageSettings
maxConcurrentConnections The max concurrent connections to connect data source at the same time. int

Constraints:
Min value = 1
recursive If true, files or sub-folders under current folder path will be deleted recursively. Default is false. Type: boolean (or Expression with resultType boolean).
storeSettings Delete activity store settings. StoreReadSettings

ExecuteDataFlowActivity

Name Description Value
type Type of activity. 'ExecuteDataFlow' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Execute data flow activity properties. ExecuteDataFlowActivityTypeProperties (required)

ExecuteDataFlowActivityTypeProperties

Name Description Value
compute Compute properties for data flow activity. ExecuteDataFlowActivityTypePropertiesCompute
continueOnError Continue on error setting used for data flow execution. Enables processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean)
dataFlow Data flow reference. DataFlowReference (required)
integrationRuntime The integration runtime reference. IntegrationRuntimeReference
runConcurrently Concurrent run setting used for data flow execution. Allows sinks with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean)
sourceStagingConcurrency Specify number of parallel staging for sources applicable to the sink. Type: integer (or Expression with resultType integer)
staging Staging info for execute data flow activity. DataFlowStagingInfo
traceLevel Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string)

ExecuteDataFlowActivityTypePropertiesCompute

Name Description Value
computeType Compute type of the cluster which will execute data flow job. Possible values include: 'General', 'MemoryOptimized', 'ComputeOptimized'. Type: string (or Expression with resultType string)
coreCount Core count of the cluster which will execute data flow job. Supported values are: 8, 16, 32, 48, 80, 144 and 272. Type: integer (or Expression with resultType integer)

DataFlowReference

Name Description Value
datasetParameters Reference data flow parameters from dataset.
parameters Data flow parameters ParameterValueSpecification
referenceName Reference data flow name. string (required)
type Data flow reference type. 'DataFlowReference' (required)
{customized property}

IntegrationRuntimeReference

Name Description Value
parameters Arguments for integration runtime. ParameterValueSpecification
referenceName Reference integration runtime name. string (required)
type Type of integration runtime. 'IntegrationRuntimeReference' (required)

DataFlowStagingInfo

Name Description Value
folderPath Folder path for staging blob. Type: string (or Expression with resultType string)
linkedService Staging linked service reference. LinkedServiceReference

ExecutePipelineActivity

Name Description Value
type Type of activity. 'ExecutePipeline' (required)
policy Execute pipeline activity policy. ExecutePipelineActivityPolicy
typeProperties Execute pipeline activity properties. ExecutePipelineActivityTypeProperties (required)

ExecutePipelineActivityPolicy

Name Description Value
secureInput When set to true, Input from activity is considered as secure and will not be logged to monitoring. bool
{customized property}

ExecutePipelineActivityTypeProperties

Name Description Value
parameters Pipeline parameters. ParameterValueSpecification
pipeline Pipeline reference. PipelineReference (required)
waitOnCompletion Defines whether activity execution will wait for the dependent pipeline execution to finish. Default is false. bool

PipelineReference

Name Description Value
name Reference name. string
referenceName Reference pipeline name. string (required)
type Pipeline reference type. 'PipelineReference' (required)

ExecuteSsisPackageActivity

Name Description Value
type Type of activity. 'ExecuteSSISPackage' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Execute SSIS package activity properties. ExecuteSsisPackageActivityTypeProperties (required)

ExecuteSsisPackageActivityTypeProperties

Name Description Value
connectVia The integration runtime reference. IntegrationRuntimeReference (required)
environmentPath The environment path to execute the SSIS package. Type: string (or Expression with resultType string).
executionCredential The package execution credential. SsisExecutionCredential
loggingLevel The logging level of SSIS package execution. Type: string (or Expression with resultType string).
logLocation SSIS package execution log location. SsisLogLocation
packageConnectionManagers The package level connection managers to execute the SSIS package. ExecuteSsisPackageActivityTypePropertiesPackageConne...
packageLocation SSIS package location. SsisPackageLocation (required)
packageParameters The package level parameters to execute the SSIS package. ExecuteSsisPackageActivityTypePropertiesPackageParam...
projectConnectionManagers The project level connection managers to execute the SSIS package. ExecuteSsisPackageActivityTypePropertiesProjectConne...
projectParameters The project level parameters to execute the SSIS package. ExecuteSsisPackageActivityTypePropertiesProjectParam...
propertyOverrides The property overrides to execute the SSIS package. ExecuteSsisPackageActivityTypePropertiesPropertyOver...
runtime Specifies the runtime to execute SSIS package. The value should be "x86" or "x64". Type: string (or Expression with resultType string).

SsisExecutionCredential

Name Description Value
domain Domain for windows authentication. Type: string (or Expression with resultType string).
password Password for windows authentication. SecureString (required)
userName UseName for windows authentication. Type: string (or Expression with resultType string).

SecureString

Name Description Value
type Type of the secret. string (required)
value Value of secure string. string (required)

SsisLogLocation

Name Description Value
logPath The SSIS package execution log path. Type: string (or Expression with resultType string).
type The type of SSIS log location. 'File' (required)
typeProperties SSIS package execution log location properties. SsisLogLocationTypeProperties (required)

SsisLogLocationTypeProperties

Name Description Value
accessCredential The package execution log access credential. SsisAccessCredential
logRefreshInterval Specifies the interval to refresh log. The default interval is 5 minutes. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

SsisAccessCredential

Name Description Value
domain Domain for windows authentication. Type: string (or Expression with resultType string).
password Password for windows authentication. SecretBase (required)
userName UseName for windows authentication. Type: string (or Expression with resultType string).

SecretBase

Name Description Value
type Set the object type AzureKeyVaultSecret
SecureString (required)

AzureKeyVaultSecretReference

Name Description Value
secretName The name of the secret in Azure Key Vault. Type: string (or Expression with resultType string).
secretVersion The version of the secret in Azure Key Vault. The default value is the latest version of the secret. Type: string (or Expression with resultType string).
store The Azure Key Vault linked service reference. LinkedServiceReference (required)
type Type of the secret. string (required)

ExecuteSsisPackageActivityTypePropertiesPackageConne...

Name Description Value
{customized property} SsisConnectionManager

SsisConnectionManager

Name Description Value
{customized property} SsisExecutionParameter

SsisExecutionParameter

Name Description Value
value SSIS package execution parameter value. Type: string (or Expression with resultType string).

SsisPackageLocation

Name Description Value
packagePath The SSIS package path. Type: string (or Expression with resultType string).
type The type of SSIS package location. 'File'
'InlinePackage'
'PackageStore'
'SSISDB'
typeProperties SSIS package location properties. SsisPackageLocationTypeProperties

SsisPackageLocationTypeProperties

Name Description Value
accessCredential The package access credential. SsisAccessCredential
childPackages The embedded child package list. SsisChildPackage[]
configurationAccessCredential The configuration file access credential. SsisAccessCredential
configurationPath The configuration file of the package execution. Type: string (or Expression with resultType string).
packageContent The embedded package content. Type: string (or Expression with resultType string).
packageLastModifiedDate The embedded package last modified date. string
packageName The package name. string
packagePassword Password of the package. SecretBase

SsisChildPackage

Name Description Value
packageContent Content for embedded child package. Type: string (or Expression with resultType string).
packageLastModifiedDate Last modified date for embedded child package. string
packageName Name for embedded child package. string
packagePath Path for embedded child package. Type: string (or Expression with resultType string).

ExecuteSsisPackageActivityTypePropertiesPackageParam...

Name Description Value
{customized property} SsisExecutionParameter

ExecuteSsisPackageActivityTypePropertiesProjectConne...

Name Description Value
{customized property} SsisConnectionManager

ExecuteSsisPackageActivityTypePropertiesProjectParam...

Name Description Value
{customized property} SsisExecutionParameter

ExecuteSsisPackageActivityTypePropertiesPropertyOver...

Name Description Value
{customized property} SsisPropertyOverride

SsisPropertyOverride

Name Description Value
isSensitive Whether SSIS package property override value is sensitive data. Value will be encrypted in SSISDB if it is true bool
value SSIS package property override value. Type: string (or Expression with resultType string).

ExecuteWranglingDataflowActivity

Name Description Value
type Type of activity. 'ExecuteWranglingDataflow' (required)
policy Activity policy. ActivityPolicy
typeProperties Execute power query activity properties. ExecutePowerQueryActivityTypeProperties (required)

ExecutePowerQueryActivityTypeProperties

Name Description Value
compute Compute properties for data flow activity. ExecuteDataFlowActivityTypePropertiesCompute
continueOnError Continue on error setting used for data flow execution. Enables processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean)
dataFlow Data flow reference. DataFlowReference (required)
integrationRuntime The integration runtime reference. IntegrationRuntimeReference
queries List of mapping for Power Query mashup query to sink dataset(s). PowerQuerySinkMapping[]
runConcurrently Concurrent run setting used for data flow execution. Allows sinks with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean)
sinks (Deprecated. Please use Queries). List of Power Query activity sinks mapped to a queryName. ExecutePowerQueryActivityTypePropertiesSinks
sourceStagingConcurrency Specify number of parallel staging for sources applicable to the sink. Type: integer (or Expression with resultType integer)
staging Staging info for execute data flow activity. DataFlowStagingInfo
traceLevel Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string)

PowerQuerySinkMapping

Name Description Value
dataflowSinks List of sinks mapped to Power Query mashup query. PowerQuerySink[]
queryName Name of the query in Power Query mashup document. string

PowerQuerySink

Name Description Value
dataset Dataset reference. DatasetReference
description Transformation description. string
flowlet Flowlet Reference DataFlowReference
linkedService Linked service reference. LinkedServiceReference
name Transformation name. string (required)
rejectedDataLinkedService Rejected data linked service reference. LinkedServiceReference
schemaLinkedService Schema linked service reference. LinkedServiceReference
script sink script. string

ExecutePowerQueryActivityTypePropertiesSinks

Name Description Value
{customized property} PowerQuerySink

FailActivity

Name Description Value
type Type of activity. 'Fail' (required)
typeProperties Fail activity properties. FailActivityTypeProperties (required)

FailActivityTypeProperties

Name Description Value
errorCode The error code that categorizes the error type of the Fail activity. It can be dynamic content that's evaluated to a non empty/blank string at runtime. Type: string (or Expression with resultType string).
message The error message that surfaced in the Fail activity. It can be dynamic content that's evaluated to a non empty/blank string at runtime. Type: string (or Expression with resultType string).

FilterActivity

Name Description Value
type Type of activity. 'Filter' (required)
typeProperties Filter activity properties. FilterActivityTypeProperties (required)

FilterActivityTypeProperties

Name Description Value
condition Condition to be used for filtering the input. Expression (required)
items Input array on which filter should be applied. Expression (required)

Expression

Name Description Value
type Expression type. 'Expression' (required)
value Expression value. string (required)

ForEachActivity

Name Description Value
type Type of activity. 'ForEach' (required)
typeProperties ForEach activity properties. ForEachActivityTypeProperties (required)

ForEachActivityTypeProperties

Name Description Value
activities List of activities to execute . Activity[] (required)
batchCount Batch count to be used for controlling the number of parallel execution (when isSequential is set to false). int

Constraints:
Max value = 50
isSequential Should the loop be executed in sequence or in parallel (max 50) bool
items Collection to iterate. Expression (required)

GetMetadataActivity

Name Description Value
type Type of activity. 'GetMetadata' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties GetMetadata activity properties. GetMetadataActivityTypeProperties (required)

GetMetadataActivityTypeProperties

Name Description Value
dataset GetMetadata activity dataset reference. DatasetReference (required)
fieldList Fields of metadata to get from dataset. any[]
formatSettings GetMetadata activity format settings. FormatReadSettings
storeSettings GetMetadata activity store settings. StoreReadSettings

FormatReadSettings

Name Description Value
type Set the object type BinaryReadSettings
DelimitedTextReadSettings
JsonReadSettings
ParquetReadSettings
XmlReadSettings (required)

HDInsightHiveActivity

Name Description Value
type Type of activity. 'HDInsightHive' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties HDInsight Hive activity properties. HDInsightHiveActivityTypeProperties (required)

HDInsightHiveActivityTypeProperties

Name Description Value
arguments User specified arguments to HDInsightActivity. any[]
defines Allows user to specify defines for Hive job request. HDInsightHiveActivityTypePropertiesDefines
getDebugInfo Debug info option. 'Always'
'Failure'
'None'
queryTimeout Query timeout value (in minutes). Effective when the HDInsight cluster is with ESP (Enterprise Security Package) int
scriptLinkedService Script linked service reference. LinkedServiceReference
scriptPath Script path. Type: string (or Expression with resultType string).
storageLinkedServices Storage linked service references. LinkedServiceReference[]
variables User specified arguments under hivevar namespace. HDInsightHiveActivityTypePropertiesVariables

HDInsightHiveActivityTypePropertiesDefines

Name Description Value
{customized property}

HDInsightHiveActivityTypePropertiesVariables

Name Description Value
{customized property}

HDInsightMapReduceActivity

Name Description Value
type Type of activity. 'HDInsightMapReduce' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties HDInsight MapReduce activity properties. HDInsightMapReduceActivityTypeProperties (required)

HDInsightMapReduceActivityTypeProperties

Name Description Value
arguments User specified arguments to HDInsightActivity. any[]
className Class name. Type: string (or Expression with resultType string).
defines Allows user to specify defines for the MapReduce job request. HDInsightMapReduceActivityTypePropertiesDefines
getDebugInfo Debug info option. 'Always'
'Failure'
'None'
jarFilePath Jar path. Type: string (or Expression with resultType string).
jarLibs Jar libs. any[]
jarLinkedService Jar linked service reference. LinkedServiceReference
storageLinkedServices Storage linked service references. LinkedServiceReference[]

HDInsightMapReduceActivityTypePropertiesDefines

Name Description Value
{customized property}

HDInsightPigActivity

Name Description Value
type Type of activity. 'HDInsightPig' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties HDInsight Pig activity properties. HDInsightPigActivityTypeProperties (required)

HDInsightPigActivityTypeProperties

Name Description Value
arguments User specified arguments to HDInsightActivity. Type: array (or Expression with resultType array).
defines Allows user to specify defines for Pig job request. HDInsightPigActivityTypePropertiesDefines
getDebugInfo Debug info option. 'Always'
'Failure'
'None'
scriptLinkedService Script linked service reference. LinkedServiceReference
scriptPath Script path. Type: string (or Expression with resultType string).
storageLinkedServices Storage linked service references. LinkedServiceReference[]

HDInsightPigActivityTypePropertiesDefines

Name Description Value
{customized property}

HDInsightSparkActivity

Name Description Value
type Type of activity. 'HDInsightSpark' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties HDInsight spark activity properties. HDInsightSparkActivityTypeProperties (required)

HDInsightSparkActivityTypeProperties

Name Description Value
arguments The user-specified arguments to HDInsightSparkActivity. any[]
className The application's Java/Spark main class. string
entryFilePath The relative path to the root folder of the code/package to be executed. Type: string (or Expression with resultType string).
getDebugInfo Debug info option. 'Always'
'Failure'
'None'
proxyUser The user to impersonate that will execute the job. Type: string (or Expression with resultType string).
rootPath The root path in 'sparkJobLinkedService' for all the job’s files. Type: string (or Expression with resultType string).
sparkConfig Spark configuration property. HDInsightSparkActivityTypePropertiesSparkConfig
sparkJobLinkedService The storage linked service for uploading the entry file and dependencies, and for receiving logs. LinkedServiceReference

HDInsightSparkActivityTypePropertiesSparkConfig

Name Description Value
{customized property}

HDInsightStreamingActivity

Name Description Value
type Type of activity. 'HDInsightStreaming' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties HDInsight streaming activity properties. HDInsightStreamingActivityTypeProperties (required)

HDInsightStreamingActivityTypeProperties

Name Description Value
arguments User specified arguments to HDInsightActivity. any[]
combiner Combiner executable name. Type: string (or Expression with resultType string).
commandEnvironment Command line environment values. any[]
defines Allows user to specify defines for streaming job request. HDInsightStreamingActivityTypePropertiesDefines
fileLinkedService Linked service reference where the files are located. LinkedServiceReference
filePaths Paths to streaming job files. Can be directories. any[] (required)
getDebugInfo Debug info option. 'Always'
'Failure'
'None'
input Input blob path. Type: string (or Expression with resultType string).
mapper Mapper executable name. Type: string (or Expression with resultType string).
output Output blob path. Type: string (or Expression with resultType string).
reducer Reducer executable name. Type: string (or Expression with resultType string).
storageLinkedServices Storage linked service references. LinkedServiceReference[]

HDInsightStreamingActivityTypePropertiesDefines

Name Description Value
{customized property}

IfConditionActivity

Name Description Value
type Type of activity. 'IfCondition' (required)
typeProperties IfCondition activity properties. IfConditionActivityTypeProperties (required)

IfConditionActivityTypeProperties

Name Description Value
expression An expression that would evaluate to Boolean. This is used to determine the block of activities (ifTrueActivities or ifFalseActivities) that will be executed. Expression (required)
ifFalseActivities List of activities to execute if expression is evaluated to false. This is an optional property and if not provided, the activity will exit without any action. Activity[]
ifTrueActivities List of activities to execute if expression is evaluated to true. This is an optional property and if not provided, the activity will exit without any action. Activity[]

LookupActivity

Name Description Value
type Type of activity. 'Lookup' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Lookup activity properties. LookupActivityTypeProperties (required)

LookupActivityTypeProperties

Name Description Value
dataset Lookup activity dataset reference. DatasetReference (required)
firstRowOnly Whether to return first row or all rows. Default value is true. Type: boolean (or Expression with resultType boolean).
source Dataset-specific source properties, same as copy activity source. CopySource (required)

ScriptActivity

Name Description Value
type Type of activity. 'Script' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Script activity properties. ScriptActivityTypeProperties (required)

ScriptActivityTypeProperties

Name Description Value
logSettings Log settings of script activity. ScriptActivityTypePropertiesLogSettings
scriptBlockExecutionTimeout ScriptBlock execution timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
scripts Array of script blocks. Type: array. ScriptActivityScriptBlock[]

ScriptActivityTypePropertiesLogSettings

Name Description Value
logDestination The destination of logs. Type: string. 'ActivityOutput'
'ExternalStore' (required)
logLocationSettings Log location settings customer needs to provide when enabling log. LogLocationSettings

ScriptActivityScriptBlock

Name Description Value
parameters Array of script parameters. Type: array. ScriptActivityParameter[]
text The query text. Type: string (or Expression with resultType string).
type The type of the query. Type: string. 'NonQuery'
'Query' (required)

ScriptActivityParameter

Name Description Value
direction The direction of the parameter. 'Input'
'InputOutput'
'Output'
name The name of the parameter. Type: string (or Expression with resultType string).
size The size of the output direction parameter. int
type The type of the parameter. 'Boolean'
'DateTime'
'DateTimeOffset'
'Decimal'
'Double'
'Guid'
'Int16'
'Int32'
'Int64'
'Single'
'String'
'Timespan'
value The value of the parameter. Type: string (or Expression with resultType string).

SetVariableActivity

Name Description Value
type Type of activity. 'SetVariable' (required)
policy Activity policy. SecureInputOutputPolicy
typeProperties Set Variable activity properties. SetVariableActivityTypeProperties (required)

SecureInputOutputPolicy

Name Description Value
secureInput When set to true, Input from activity is considered as secure and will not be logged to monitoring. bool
secureOutput When set to true, Output from activity is considered as secure and will not be logged to monitoring. bool

SetVariableActivityTypeProperties

Name Description Value
setSystemVariable If set to true, it sets the pipeline run return value. bool
value Value to be set. Could be a static value or Expression.
variableName Name of the variable whose value needs to be set. string

SynapseSparkJobDefinitionActivity

Name Description Value
type Type of activity. 'SparkJob' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Execute spark job activity properties. SynapseSparkJobActivityTypeProperties (required)

SynapseSparkJobActivityTypeProperties

Name Description Value
args User specified arguments to SynapseSparkJobDefinitionActivity. any[]
className The fully-qualified identifier or the main class that is in the main definition file, which will override the 'className' of the spark job definition you provide. Type: string (or Expression with resultType string).
conf Spark configuration properties, which will override the 'conf' of the spark job definition you provide.
configurationType The type of the spark config. 'Artifact'
'Customized'
'Default'
driverSize Number of core and memory to be used for driver allocated in the specified Spark pool for the job, which will be used for overriding 'driverCores' and 'driverMemory' of the spark job definition you provide. Type: string (or Expression with resultType string).
executorSize Number of core and memory to be used for executors allocated in the specified Spark pool for the job, which will be used for overriding 'executorCores' and 'executorMemory' of the spark job definition you provide. Type: string (or Expression with resultType string).
file The main file used for the job, which will override the 'file' of the spark job definition you provide. Type: string (or Expression with resultType string).
files (Deprecated. Please use pythonCodeReference and filesV2) Additional files used for reference in the main definition file, which will override the 'files' of the spark job definition you provide. any[]
filesV2 Additional files used for reference in the main definition file, which will override the 'jars' and 'files' of the spark job definition you provide. any[]
numExecutors Number of executors to launch for this job, which will override the 'numExecutors' of the spark job definition you provide. Type: integer (or Expression with resultType integer).
pythonCodeReference Additional python code files used for reference in the main definition file, which will override the 'pyFiles' of the spark job definition you provide. any[]
scanFolder Scanning subfolders from the root folder of the main definition file, these files will be added as reference files. The folders named 'jars', 'pyFiles', 'files' or 'archives' will be scanned, and the folders name are case sensitive. Type: boolean (or Expression with resultType boolean).
sparkConfig Spark configuration property. SynapseSparkJobActivityTypePropertiesSparkConfig
sparkJob Synapse spark job reference. SynapseSparkJobReference (required)
targetBigDataPool The name of the big data pool which will be used to execute the spark batch job, which will override the 'targetBigDataPool' of the spark job definition you provide. BigDataPoolParametrizationReference
targetSparkConfiguration The spark configuration of the spark job. SparkConfigurationParametrizationReference

SynapseSparkJobActivityTypePropertiesSparkConfig

Name Description Value
{customized property}

SynapseSparkJobReference

Name Description Value
referenceName Reference spark job name. Expression with resultType string.
type Synapse spark job reference type. 'SparkJobDefinitionReference' (required)

BigDataPoolParametrizationReference

Name Description Value
referenceName Reference big data pool name. Type: string (or Expression with resultType string).
type Big data pool reference type. 'BigDataPoolReference' (required)

SparkConfigurationParametrizationReference

Name Description Value
referenceName Reference spark configuration name. Type: string (or Expression with resultType string).
type Spark configuration reference type. 'SparkConfigurationReference' (required)

SqlServerStoredProcedureActivity

Name Description Value
type Type of activity. 'SqlServerStoredProcedure' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties SQL stored procedure activity properties. SqlServerStoredProcedureActivityTypeProperties (required)

SqlServerStoredProcedureActivityTypeProperties

Name Description Value
storedProcedureName Stored procedure name. Type: string (or Expression with resultType string).
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}".

SwitchActivity

Name Description Value
type Type of activity. 'Switch' (required)
typeProperties Switch activity properties. SwitchActivityTypeProperties (required)

SwitchActivityTypeProperties

Name Description Value
cases List of cases that correspond to expected values of the 'on' property. This is an optional property and if not provided, the activity will execute activities provided in defaultActivities. SwitchCase[]
defaultActivities List of activities to execute if no case condition is satisfied. This is an optional property and if not provided, the activity will exit without any action. Activity[]
on An expression that would evaluate to a string or integer. This is used to determine the block of activities in cases that will be executed. Expression (required)

SwitchCase

Name Description Value
activities List of activities to execute for satisfied case condition. Activity[]
value Expected value that satisfies the expression result of the 'on' property. string

SynapseNotebookActivity

Name Description Value
type Type of activity. 'SynapseNotebook' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Execute Synapse notebook activity properties. SynapseNotebookActivityTypeProperties (required)

SynapseNotebookActivityTypeProperties

Name Description Value
conf Spark configuration properties, which will override the 'conf' of the notebook you provide.
configurationType The type of the spark config. 'Artifact'
'Customized'
'Default'
driverSize Number of core and memory to be used for driver allocated in the specified Spark pool for the session, which will be used for overriding 'driverCores' and 'driverMemory' of the notebook you provide. Type: string (or Expression with resultType string).
executorSize Number of core and memory to be used for executors allocated in the specified Spark pool for the session, which will be used for overriding 'executorCores' and 'executorMemory' of the notebook you provide. Type: string (or Expression with resultType string).
notebook Synapse notebook reference. SynapseNotebookReference (required)
numExecutors Number of executors to launch for this session, which will override the 'numExecutors' of the notebook you provide. Type: integer (or Expression with resultType integer).
parameters Notebook parameters. SynapseNotebookActivityTypePropertiesParameters
sparkConfig Spark configuration property. SynapseNotebookActivityTypePropertiesSparkConfig
sparkPool The name of the big data pool which will be used to execute the notebook. BigDataPoolParametrizationReference
targetSparkConfiguration The spark configuration of the spark job. SparkConfigurationParametrizationReference

SynapseNotebookReference

Name Description Value
referenceName Reference notebook name. Type: string (or Expression with resultType string).
type Synapse notebook reference type. 'NotebookReference' (required)

SynapseNotebookActivityTypePropertiesParameters

Name Description Value
{customized property} NotebookParameter

NotebookParameter

Name Description Value
type Notebook parameter type. 'bool'
'float'
'int'
'string'
value Notebook parameter value. Type: string (or Expression with resultType string).

SynapseNotebookActivityTypePropertiesSparkConfig

Name Description Value
{customized property}

UntilActivity

Name Description Value
type Type of activity. 'Until' (required)
typeProperties Until activity properties. UntilActivityTypeProperties (required)

UntilActivityTypeProperties

Name Description Value
activities List of activities to execute. Activity[] (required)
expression An expression that would evaluate to Boolean. The loop will continue until this expression evaluates to true Expression (required)
timeout Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

ValidationActivity

Name Description Value
type Type of activity. 'Validation' (required)
typeProperties Validation activity properties. ValidationActivityTypeProperties (required)

ValidationActivityTypeProperties

Name Description Value
childItems Can be used if dataset points to a folder. If set to true, the folder must have at least one file. If set to false, the folder must be empty. Type: boolean (or Expression with resultType boolean).
dataset Validation activity dataset reference. DatasetReference (required)
minimumSize Can be used if dataset points to a file. The file must be greater than or equal in size to the value specified. Type: integer (or Expression with resultType integer).
sleep A delay in seconds between validation attempts. If no value is specified, 10 seconds will be used as the default. Type: integer (or Expression with resultType integer).
timeout Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

WaitActivity

Name Description Value
type Type of activity. 'Wait' (required)
typeProperties Wait activity properties. WaitActivityTypeProperties (required)

WaitActivityTypeProperties

Name Description Value
waitTimeInSeconds Duration in seconds. Type: integer (or Expression with resultType integer).

WebActivity

Name Description Value
type Type of activity. 'WebActivity' (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Web activity properties. WebActivityTypeProperties (required)

WebActivityTypeProperties

Name Description Value
authentication Authentication method used for calling the endpoint. WebActivityAuthentication
body Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string).
connectVia The integration runtime reference. IntegrationRuntimeReference
datasets List of datasets passed to web endpoint. DatasetReference[]
disableCertValidation When set to true, Certificate validation will be disabled. bool
headers Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). WebActivityTypePropertiesHeaders
httpRequestTimeout Timeout for the HTTP request to get a response. Format is in TimeSpan (hh:mm:ss). This value is the timeout to get a response, not the activity timeout. The default value is 00:01:00 (1 minute). The range is from 1 to 10 minutes
linkedServices List of linked services passed to web endpoint. LinkedServiceReference[]
method Rest API method for target endpoint. 'DELETE'
'GET'
'POST'
'PUT' (required)
turnOffAsync Option to disable invoking HTTP GET on location given in response header of a HTTP 202 Response. If set true, it stops invoking HTTP GET on http location given in response header. If set false then continues to invoke HTTP GET call on location given in http response headers. bool
url Web activity target endpoint and path. Type: string (or Expression with resultType string).

WebActivityAuthentication

Name Description Value
credential The credential reference containing authentication information. CredentialReference
password Password for the PFX file or basic authentication / Secret when used for ServicePrincipal SecretBase
pfx Base64-encoded contents of a PFX file or Certificate when used for ServicePrincipal SecretBase
resource Resource for which Azure Auth token will be requested when using MSI Authentication. Type: string (or Expression with resultType string).
type Web activity authentication (Basic/ClientCertificate/MSI/ServicePrincipal) string
username Web activity authentication user name for basic authentication or ClientID when used for ServicePrincipal. Type: string (or Expression with resultType string).
userTenant TenantId for which Azure Auth token will be requested when using ServicePrincipal Authentication. Type: string (or Expression with resultType string).

CredentialReference

Name Description Value
referenceName Reference credential name. string (required)
type Credential reference type. 'CredentialReference' (required)
{customized property}

WebActivityTypePropertiesHeaders

Name Description Value
{customized property} string

WebHookActivity

Name Description Value
type Type of activity. 'WebHook' (required)
policy Activity policy. SecureInputOutputPolicy
typeProperties WebHook activity properties. WebHookActivityTypeProperties (required)

WebHookActivityTypeProperties

Name Description Value
authentication Authentication method used for calling the endpoint. WebActivityAuthentication
body Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string).
headers Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). WebHookActivityTypePropertiesHeaders
method Rest API method for target endpoint. 'POST' (required)
reportStatusOnCallBack When set to true, statusCode, output and error in callback request body will be consumed by activity. The activity can be marked as failed by setting statusCode >= 400 in callback request. Default is false. Type: boolean (or Expression with resultType boolean).
timeout The timeout within which the webhook should be called back. If there is no value specified, it defaults to 10 minutes. Type: string. Pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). string
url WebHook activity target endpoint and path. Type: string (or Expression with resultType string).

WebHookActivityTypePropertiesHeaders

Name Description Value
{customized property} string

PipelineFolder

Name Description Value
name The name of the folder that this Pipeline is in. string

ParameterDefinitionSpecification

Name Description Value
{customized property} ParameterSpecification

ParameterSpecification

Name Description Value
defaultValue Default value of parameter.
type Parameter type. 'Array'
'Bool'
'Float'
'Int'
'Object'
'SecureString'
'String' (required)

PipelinePolicy

Name Description Value
elapsedTimeMetric Pipeline ElapsedTime Metric Policy. PipelineElapsedTimeMetricPolicy

PipelineElapsedTimeMetricPolicy

Name Description Value
duration TimeSpan value, after which an Azure Monitoring Metric is fired.

PipelineRunDimensions

Name Description Value
{customized property}

VariableDefinitionSpecification

Name Description Value
{customized property} VariableSpecification

VariableSpecification

Name Description Value
defaultValue Default value of variable.
type Variable type. 'Array'
'Bool'
'String' (required)

Quickstart templates

The following quickstart templates deploy this resource type.

Template Description
More is possible with Azure Data Factory - One click to try Azure Data Factory

Deploy to Azure
This template creates a data factory pipeline for a copy activity from Azure Blob into another Azure Blob
Create a Data Factory, copy from Salesforce to Blob

Deploy to Azure
This template uses Azure Data Factory to perform data movement from Salesforce to Azure Blobs
Bulk copy using Azure Data Factory

Deploy to Azure
This template creates a V2 data factory that copies data in bulk an Azure SQL database to an Azure SQL data warehouse.
Create a V2 data factory

Deploy to Azure
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to another folder in the storage.
Create a V2 data factory (MySQL)

Deploy to Azure
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to a table in an Azure Database for MySQL.
Create a V2 data factory (PostGre)

Deploy to Azure
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to a table in an Azure Database for PostgreSQL.
Create a V2 data factory (SQL)

Deploy to Azure
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to a table in an Azure SQL database.
Create a V2 data factory (SQL On-prem)

Deploy to Azure
This template creates a V2 data factory that copies data from an on-premises SQL Server to an Azure blob storage
Create a V2 data factory (Spark)

Deploy to Azure
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to another folder in the storage.

Terraform (AzAPI provider) resource definition

The factories/pipelines resource type can be deployed with operations that target:

  • Resource groups

For a list of changed properties in each API version, see change log.

Resource format

To create a Microsoft.DataFactory/factories/pipelines resource, add the following Terraform to your template.

resource "azapi_resource" "symbolicname" {
  type = "Microsoft.DataFactory/factories/pipelines@2018-06-01"
  name = "string"
  parent_id = "string"
  body = jsonencode({
    properties = {
      activities = [
        {
          dependsOn = [
            {
              activity = "string"
              dependencyConditions = [
                "string"
              ]
            }
          ]
          description = "string"
          name = "string"
          onInactiveMarkAs = "string"
          state = "string"
          userProperties = [
            {
              name = "string"
            }
          ]
          type = "string"
          // For remaining properties, see Activity objects
        }
      ]
      annotations = [ object ]
      concurrency = int
      description = "string"
      folder = {
        name = "string"
      }
      parameters = {
        {customized property} = {
          type = "string"
        }
      }
      policy = {
        elapsedTimeMetric = {}
      }
      runDimensions = {}
      variables = {
        {customized property} = {
          type = "string"
        }
      }
    }
  })
}

Activity objects

Set the type property to specify the type of object.

For AppendVariable, use:

  type = "AppendVariable"
  typeProperties = {
    variableName = "string"
  }

For AzureDataExplorerCommand, use:

  type = "AzureDataExplorerCommand"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {}

For AzureFunctionActivity, use:

  type = "AzureFunctionActivity"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    headers = {
      {customized property} = "string"
    }
    method = "string"
  }

For AzureMLBatchExecution, use:

  type = "AzureMLBatchExecution"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    globalParameters = {}
    webServiceInputs = {
      {customized property} = {
        linkedServiceName = {
          parameters = {}
          referenceName = "string"
          type = "LinkedServiceReference"
        }
      }
    }
    webServiceOutputs = {
      {customized property} = {
        linkedServiceName = {
          parameters = {}
          referenceName = "string"
          type = "LinkedServiceReference"
        }
      }
    }
  }

For AzureMLExecutePipeline, use:

  type = "AzureMLExecutePipeline"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {}

For AzureMLUpdateResource, use:

  type = "AzureMLUpdateResource"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    trainedModelLinkedServiceName = {
      parameters = {}
      referenceName = "string"
      type = "LinkedServiceReference"
    }
  }

For Copy, use:

  type = "Copy"
  inputs = [
    {
      parameters = {}
      referenceName = "string"
      type = "DatasetReference"
    }
  ]
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  outputs = [
    {
      parameters = {}
      referenceName = "string"
      type = "DatasetReference"
    }
  ]
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    logSettings = {
      copyActivityLogSettings = {}
      logLocationSettings = {
        linkedServiceName = {
          parameters = {}
          referenceName = "string"
          type = "LinkedServiceReference"
        }
      }
    }
    logStorageSettings = {
      linkedServiceName = {
        parameters = {}
        referenceName = "string"
        type = "LinkedServiceReference"
      }
    }
    preserve = [ object ]
    preserveRules = [ object ]
    redirectIncompatibleRowSettings = {}
    sink = {
      type = "string"
      // For remaining properties, see CopySink objects
    }
    skipErrorFile = {}
    source = {
      type = "string"
      // For remaining properties, see CopySource objects
    }
    stagingSettings = {
      linkedServiceName = {
        parameters = {}
        referenceName = "string"
        type = "LinkedServiceReference"
      }
    }
  }

For Custom, use:

  type = "Custom"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    extendedProperties = {}
    referenceObjects = {
      datasets = [
        {
          parameters = {}
          referenceName = "string"
          type = "DatasetReference"
        }
      ]
      linkedServices = [
        {
          parameters = {}
          referenceName = "string"
          type = "LinkedServiceReference"
        }
      ]
    }
    resourceLinkedService = {
      parameters = {}
      referenceName = "string"
      type = "LinkedServiceReference"
    }
  }

For DatabricksNotebook, use:

  type = "DatabricksNotebook"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    baseParameters = {}
    libraries = [
      {
      }
    ]
  }

For DatabricksSparkJar, use:

  type = "DatabricksSparkJar"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    libraries = [
      {
      }
    ]
    parameters = [ object ]
  }

For DatabricksSparkPython, use:

  type = "DatabricksSparkPython"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    libraries = [
      {
      }
    ]
    parameters = [ object ]
  }

For DataLakeAnalyticsU-SQL, use:

  type = "DataLakeAnalyticsU-SQL"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    parameters = {}
    scriptLinkedService = {
      parameters = {}
      referenceName = "string"
      type = "LinkedServiceReference"
    }
  }

For Delete, use:

  type = "Delete"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    dataset = {
      parameters = {}
      referenceName = "string"
      type = "DatasetReference"
    }
    logStorageSettings = {
      linkedServiceName = {
        parameters = {}
        referenceName = "string"
        type = "LinkedServiceReference"
      }
    }
    maxConcurrentConnections = int
    storeSettings = {
      type = "string"
      // For remaining properties, see StoreReadSettings objects
    }
  }

For ExecuteDataFlow, use:

  type = "ExecuteDataFlow"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    compute = {}
    dataFlow = {
      parameters = {}
      referenceName = "string"
      type = "DataFlowReference"
    }
    integrationRuntime = {
      parameters = {}
      referenceName = "string"
      type = "IntegrationRuntimeReference"
    }
    staging = {
      linkedService = {
        parameters = {}
        referenceName = "string"
        type = "LinkedServiceReference"
      }
    }
  }

For ExecutePipeline, use:

  type = "ExecutePipeline"
  policy = {
    secureInput = bool
  }
  typeProperties = {
    parameters = {}
    pipeline = {
      name = "string"
      referenceName = "string"
      type = "PipelineReference"
    }
    waitOnCompletion = bool
  }

For ExecuteSSISPackage, use:

  type = "ExecuteSSISPackage"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    connectVia = {
      parameters = {}
      referenceName = "string"
      type = "IntegrationRuntimeReference"
    }
    executionCredential = {
      password = {
        type = "string"
        value = "string"
      }
    }
    logLocation = {
      type = "File"
      typeProperties = {
        accessCredential = {
          password = {
            type = "string"
            // For remaining properties, see SecretBase objects
          }
        }
      }
    }
    packageConnectionManagers = {
      {customized property} = {
        {customized property} = {}
      }
    }
    packageLocation = {
      type = "string"
      typeProperties = {
        accessCredential = {
          password = {
            type = "string"
            // For remaining properties, see SecretBase objects
          }
        }
        childPackages = [
          {
            packageLastModifiedDate = "string"
            packageName = "string"
          }
        ]
        configurationAccessCredential = {
          password = {
            type = "string"
            // For remaining properties, see SecretBase objects
          }
        }
        packageLastModifiedDate = "string"
        packageName = "string"
        packagePassword = {
          type = "string"
          // For remaining properties, see SecretBase objects
        }
      }
    }
    packageParameters = {
      {customized property} = {}
    }
    projectConnectionManagers = {
      {customized property} = {
        {customized property} = {}
      }
    }
    projectParameters = {
      {customized property} = {}
    }
    propertyOverrides = {
      {customized property} = {
        isSensitive = bool
      }
    }
  }

For ExecuteWranglingDataflow, use:

  type = "ExecuteWranglingDataflow"
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    compute = {}
    dataFlow = {
      parameters = {}
      referenceName = "string"
      type = "DataFlowReference"
    }
    integrationRuntime = {
      parameters = {}
      referenceName = "string"
      type = "IntegrationRuntimeReference"
    }
    queries = [
      {
        dataflowSinks = [
          {
            dataset = {
              parameters = {}
              referenceName = "string"
              type = "DatasetReference"
            }
            description = "string"
            flowlet = {
              parameters = {}
              referenceName = "string"
              type = "DataFlowReference"
            }
            linkedService = {
              parameters = {}
              referenceName = "string"
              type = "LinkedServiceReference"
            }
            name = "string"
            rejectedDataLinkedService = {
              parameters = {}
              referenceName = "string"
              type = "LinkedServiceReference"
            }
            schemaLinkedService = {
              parameters = {}
              referenceName = "string"
              type = "LinkedServiceReference"
            }
            script = "string"
          }
        ]
        queryName = "string"
      }
    ]
    sinks = {
      {customized property} = {
        dataset = {
          parameters = {}
          referenceName = "string"
          type = "DatasetReference"
        }
        description = "string"
        flowlet = {
          parameters = {}
          referenceName = "string"
          type = "DataFlowReference"
        }
        linkedService = {
          parameters = {}
          referenceName = "string"
          type = "LinkedServiceReference"
        }
        name = "string"
        rejectedDataLinkedService = {
          parameters = {}
          referenceName = "string"
          type = "LinkedServiceReference"
        }
        schemaLinkedService = {
          parameters = {}
          referenceName = "string"
          type = "LinkedServiceReference"
        }
        script = "string"
      }
    }
    staging = {
      linkedService = {
        parameters = {}
        referenceName = "string"
        type = "LinkedServiceReference"
      }
    }
  }

For Fail, use:

  type = "Fail"
  typeProperties = {}

For Filter, use:

  type = "Filter"
  typeProperties = {
    condition = {
      type = "Expression"
      value = "string"
    }
    items = {
      type = "Expression"
      value = "string"
    }
  }

For ForEach, use:

  type = "ForEach"
  typeProperties = {
    activities = [
      {
        dependsOn = [
          {
            activity = "string"
            dependencyConditions = [
              "string"
            ]
          }
        ]
        description = "string"
        name = "string"
        onInactiveMarkAs = "string"
        state = "string"
        userProperties = [
          {
            name = "string"
          }
        ]
        type = "string"
        // For remaining properties, see Activity objects
      }
    ]
    batchCount = int
    isSequential = bool
    items = {
      type = "Expression"
      value = "string"
    }
  }

For GetMetadata, use:

  type = "GetMetadata"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    dataset = {
      parameters = {}
      referenceName = "string"
      type = "DatasetReference"
    }
    fieldList = [ object ]
    formatSettings = {
      type = "string"
      // For remaining properties, see FormatReadSettings objects
    }
    storeSettings = {
      type = "string"
      // For remaining properties, see StoreReadSettings objects
    }
  }

For HDInsightHive, use:

  type = "HDInsightHive"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    arguments = [ object ]
    defines = {}
    getDebugInfo = "string"
    queryTimeout = int
    scriptLinkedService = {
      parameters = {}
      referenceName = "string"
      type = "LinkedServiceReference"
    }
    storageLinkedServices = [
      {
        parameters = {}
        referenceName = "string"
        type = "LinkedServiceReference"
      }
    ]
    variables = {}
  }

For HDInsightMapReduce, use:

  type = "HDInsightMapReduce"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    arguments = [ object ]
    defines = {}
    getDebugInfo = "string"
    jarLibs = [ object ]
    jarLinkedService = {
      parameters = {}
      referenceName = "string"
      type = "LinkedServiceReference"
    }
    storageLinkedServices = [
      {
        parameters = {}
        referenceName = "string"
        type = "LinkedServiceReference"
      }
    ]
  }

For HDInsightPig, use:

  type = "HDInsightPig"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    defines = {}
    getDebugInfo = "string"
    scriptLinkedService = {
      parameters = {}
      referenceName = "string"
      type = "LinkedServiceReference"
    }
    storageLinkedServices = [
      {
        parameters = {}
        referenceName = "string"
        type = "LinkedServiceReference"
      }
    ]
  }

For HDInsightSpark, use:

  type = "HDInsightSpark"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    arguments = [ object ]
    className = "string"
    getDebugInfo = "string"
    sparkConfig = {}
    sparkJobLinkedService = {
      parameters = {}
      referenceName = "string"
      type = "LinkedServiceReference"
    }
  }

For HDInsightStreaming, use:

  type = "HDInsightStreaming"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    arguments = [ object ]
    commandEnvironment = [ object ]
    defines = {}
    fileLinkedService = {
      parameters = {}
      referenceName = "string"
      type = "LinkedServiceReference"
    }
    filePaths = [ object ]
    getDebugInfo = "string"
    storageLinkedServices = [
      {
        parameters = {}
        referenceName = "string"
        type = "LinkedServiceReference"
      }
    ]
  }

For IfCondition, use:

  type = "IfCondition"
  typeProperties = {
    expression = {
      type = "Expression"
      value = "string"
    }
    ifFalseActivities = [
      {
        dependsOn = [
          {
            activity = "string"
            dependencyConditions = [
              "string"
            ]
          }
        ]
        description = "string"
        name = "string"
        onInactiveMarkAs = "string"
        state = "string"
        userProperties = [
          {
            name = "string"
          }
        ]
        type = "string"
        // For remaining properties, see Activity objects
      }
    ]
    ifTrueActivities = [
      {
        dependsOn = [
          {
            activity = "string"
            dependencyConditions = [
              "string"
            ]
          }
        ]
        description = "string"
        name = "string"
        onInactiveMarkAs = "string"
        state = "string"
        userProperties = [
          {
            name = "string"
          }
        ]
        type = "string"
        // For remaining properties, see Activity objects
      }
    ]
  }

For Lookup, use:

  type = "Lookup"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    dataset = {
      parameters = {}
      referenceName = "string"
      type = "DatasetReference"
    }
    source = {
      type = "string"
      // For remaining properties, see CopySource objects
    }
  }

For Script, use:

  type = "Script"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    logSettings = {
      logDestination = "string"
      logLocationSettings = {
        linkedServiceName = {
          parameters = {}
          referenceName = "string"
          type = "LinkedServiceReference"
        }
      }
    }
    scripts = [
      {
        parameters = [
          {
            direction = "string"
            size = int
            type = "string"
          }
        ]
        type = "string"
      }
    ]
  }

For SetVariable, use:

  type = "SetVariable"
  policy = {
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    setSystemVariable = bool
    variableName = "string"
  }

For SparkJob, use:

  type = "SparkJob"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    args = [ object ]
    configurationType = "string"
    files = [ object ]
    filesV2 = [ object ]
    pythonCodeReference = [ object ]
    sparkConfig = {}
    sparkJob = {
      type = "SparkJobDefinitionReference"
    }
    targetBigDataPool = {
      type = "BigDataPoolReference"
    }
    targetSparkConfiguration = {
      type = "SparkConfigurationReference"
    }
  }

For SqlServerStoredProcedure, use:

  type = "SqlServerStoredProcedure"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {}

For Switch, use:

  type = "Switch"
  typeProperties = {
    cases = [
      {
        activities = [
          {
            dependsOn = [
              {
                activity = "string"
                dependencyConditions = [
                  "string"
                ]
              }
            ]
            description = "string"
            name = "string"
            onInactiveMarkAs = "string"
            state = "string"
            userProperties = [
              {
                name = "string"
              }
            ]
            type = "string"
            // For remaining properties, see Activity objects
          }
        ]
        value = "string"
      }
    ]
    defaultActivities = [
      {
        dependsOn = [
          {
            activity = "string"
            dependencyConditions = [
              "string"
            ]
          }
        ]
        description = "string"
        name = "string"
        onInactiveMarkAs = "string"
        state = "string"
        userProperties = [
          {
            name = "string"
          }
        ]
        type = "string"
        // For remaining properties, see Activity objects
      }
    ]
    on = {
      type = "Expression"
      value = "string"
    }
  }

For SynapseNotebook, use:

  type = "SynapseNotebook"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    configurationType = "string"
    notebook = {
      type = "NotebookReference"
    }
    parameters = {
      {customized property} = {
        type = "string"
      }
    }
    sparkConfig = {}
    sparkPool = {
      type = "BigDataPoolReference"
    }
    targetSparkConfiguration = {
      type = "SparkConfigurationReference"
    }
  }

For Until, use:

  type = "Until"
  typeProperties = {
    activities = [
      {
        dependsOn = [
          {
            activity = "string"
            dependencyConditions = [
              "string"
            ]
          }
        ]
        description = "string"
        name = "string"
        onInactiveMarkAs = "string"
        state = "string"
        userProperties = [
          {
            name = "string"
          }
        ]
        type = "string"
        // For remaining properties, see Activity objects
      }
    ]
    expression = {
      type = "Expression"
      value = "string"
    }
  }

For Validation, use:

  type = "Validation"
  typeProperties = {
    dataset = {
      parameters = {}
      referenceName = "string"
      type = "DatasetReference"
    }
  }

For Wait, use:

  type = "Wait"
  typeProperties = {}

For WebActivity, use:

  type = "WebActivity"
  linkedServiceName = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }
  policy = {
    retryIntervalInSeconds = int
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    authentication = {
      credential = {
        referenceName = "string"
        type = "CredentialReference"
      }
      password = {
        type = "string"
        // For remaining properties, see SecretBase objects
      }
      pfx = {
        type = "string"
        // For remaining properties, see SecretBase objects
      }
      type = "string"
    }
    connectVia = {
      parameters = {}
      referenceName = "string"
      type = "IntegrationRuntimeReference"
    }
    datasets = [
      {
        parameters = {}
        referenceName = "string"
        type = "DatasetReference"
      }
    ]
    disableCertValidation = bool
    headers = {
      {customized property} = "string"
    }
    linkedServices = [
      {
        parameters = {}
        referenceName = "string"
        type = "LinkedServiceReference"
      }
    ]
    method = "string"
    turnOffAsync = bool
  }

For WebHook, use:

  type = "WebHook"
  policy = {
    secureInput = bool
    secureOutput = bool
  }
  typeProperties = {
    authentication = {
      credential = {
        referenceName = "string"
        type = "CredentialReference"
      }
      password = {
        type = "string"
        // For remaining properties, see SecretBase objects
      }
      pfx = {
        type = "string"
        // For remaining properties, see SecretBase objects
      }
      type = "string"
    }
    headers = {
      {customized property} = "string"
    }
    method = "POST"
    timeout = "string"
  }

CopySink objects

Set the type property to specify the type of object.

For AvroSink, use:

  type = "AvroSink"
  formatSettings = {
    recordName = "string"
    recordNamespace = "string"
    type = "string"
  }
  storeSettings = {
    metadata = [
      {
      }
    ]
    type = "string"
    // For remaining properties, see StoreWriteSettings objects
  }

For AzureBlobFSSink, use:

  type = "AzureBlobFSSink"
  metadata = [
    {
    }
  ]

For AzureDatabricksDeltaLakeSink, use:

  type = "AzureDatabricksDeltaLakeSink"
  importSettings = {
    type = "string"
  }

For AzureDataExplorerSink, use:

  type = "AzureDataExplorerSink"

For AzureDataLakeStoreSink, use:

  type = "AzureDataLakeStoreSink"

For AzureMySqlSink, use:

  type = "AzureMySqlSink"

For AzurePostgreSqlSink, use:

  type = "AzurePostgreSqlSink"

For AzureQueueSink, use:

  type = "AzureQueueSink"

For AzureSearchIndexSink, use:

  type = "AzureSearchIndexSink"
  writeBehavior = "string"

For AzureSqlSink, use:

  type = "AzureSqlSink"
  upsertSettings = {}

For AzureTableSink, use:

  type = "AzureTableSink"

For BinarySink, use:

  type = "BinarySink"
  storeSettings = {
    metadata = [
      {
      }
    ]
    type = "string"
    // For remaining properties, see StoreWriteSettings objects
  }

For BlobSink, use:

  type = "BlobSink"
  metadata = [
    {
    }
  ]

For CommonDataServiceForAppsSink, use:

  type = "CommonDataServiceForAppsSink"
  writeBehavior = "Upsert"

For CosmosDbMongoDbApiSink, use:

  type = "CosmosDbMongoDbApiSink"

For CosmosDbSqlApiSink, use:

  type = "CosmosDbSqlApiSink"

For DelimitedTextSink, use:

  type = "DelimitedTextSink"
  formatSettings = {
    type = "string"
  }
  storeSettings = {
    metadata = [
      {
      }
    ]
    type = "string"
    // For remaining properties, see StoreWriteSettings objects
  }

For DocumentDbCollectionSink, use:

  type = "DocumentDbCollectionSink"

For DynamicsCrmSink, use:

  type = "DynamicsCrmSink"
  writeBehavior = "Upsert"

For DynamicsSink, use:

  type = "DynamicsSink"
  writeBehavior = "Upsert"

For FileSystemSink, use:

  type = "FileSystemSink"

For InformixSink, use:

  type = "InformixSink"

For JsonSink, use:

  type = "JsonSink"
  formatSettings = {
    type = "string"
  }
  storeSettings = {
    metadata = [
      {
      }
    ]
    type = "string"
    // For remaining properties, see StoreWriteSettings objects
  }

For LakeHouseTableSink, use:

  type = "LakeHouseTableSink"

For MicrosoftAccessSink, use:

  type = "MicrosoftAccessSink"

For MongoDbAtlasSink, use:

  type = "MongoDbAtlasSink"

For MongoDbV2Sink, use:

  type = "MongoDbV2Sink"

For OdbcSink, use:

  type = "OdbcSink"

For OracleSink, use:

  type = "OracleSink"

For OrcSink, use:

  type = "OrcSink"
  formatSettings = {
    type = "string"
  }
  storeSettings = {
    metadata = [
      {
      }
    ]
    type = "string"
    // For remaining properties, see StoreWriteSettings objects
  }

For ParquetSink, use:

  type = "ParquetSink"
  formatSettings = {
    type = "string"
  }
  storeSettings = {
    metadata = [
      {
      }
    ]
    type = "string"
    // For remaining properties, see StoreWriteSettings objects
  }

For RestSink, use:

  type = "RestSink"

For SalesforceServiceCloudSink, use:

  type = "SalesforceServiceCloudSink"
  writeBehavior = "string"

For SalesforceServiceCloudV2Sink, use:

  type = "SalesforceServiceCloudV2Sink"
  writeBehavior = "string"

For SalesforceSink, use:

  type = "SalesforceSink"
  writeBehavior = "string"

For SalesforceV2Sink, use:

  type = "SalesforceV2Sink"
  writeBehavior = "string"

For SapCloudForCustomerSink, use:

  type = "SapCloudForCustomerSink"
  writeBehavior = "string"

For SnowflakeSink, use:

  type = "SnowflakeSink"
  importSettings = {
    additionalCopyOptions = {}
    additionalFormatOptions = {}
    type = "string"
  }

For SnowflakeV2Sink, use:

  type = "SnowflakeV2Sink"
  importSettings = {
    additionalCopyOptions = {}
    additionalFormatOptions = {}
    type = "string"
  }

For SqlDWSink, use:

  type = "SqlDWSink"
  copyCommandSettings = {
    additionalOptions = {
      {customized property} = "string"
    }
    defaultValues = [
      {
      }
    ]
  }
  polyBaseSettings = {
    rejectType = "string"
  }
  upsertSettings = {}

For SqlMISink, use:

  type = "SqlMISink"
  upsertSettings = {}

For SqlServerSink, use:

  type = "SqlServerSink"
  upsertSettings = {}

For SqlSink, use:

  type = "SqlSink"
  upsertSettings = {}

For WarehouseSink, use:

  type = "WarehouseSink"
  copyCommandSettings = {
    additionalOptions = {
      {customized property} = "string"
    }
    defaultValues = [
      {
      }
    ]
  }

StoreWriteSettings objects

Set the type property to specify the type of object.

For AzureBlobFSWriteSettings, use:

  type = "AzureBlobFSWriteSettings"

For AzureBlobStorageWriteSettings, use:

  type = "AzureBlobStorageWriteSettings"

For AzureDataLakeStoreWriteSettings, use:

  type = "AzureDataLakeStoreWriteSettings"

For AzureFileStorageWriteSettings, use:

  type = "AzureFileStorageWriteSettings"

For FileServerWriteSettings, use:

  type = "FileServerWriteSettings"

For LakeHouseWriteSettings, use:

  type = "LakeHouseWriteSettings"

For SftpWriteSettings, use:

  type = "SftpWriteSettings"

CopySource objects

Set the type property to specify the type of object.

For AmazonMWSSource, use:

  type = "AmazonMWSSource"

For AmazonRdsForOracleSource, use:

  type = "AmazonRdsForOracleSource"
  partitionSettings = {}

For AmazonRdsForSqlServerSource, use:

  type = "AmazonRdsForSqlServerSource"
  partitionSettings = {}

For AmazonRedshiftSource, use:

  type = "AmazonRedshiftSource"
  redshiftUnloadSettings = {
    s3LinkedServiceName = {
      parameters = {}
      referenceName = "string"
      type = "LinkedServiceReference"
    }
  }

For AvroSource, use:

  type = "AvroSource"
  storeSettings = {
    type = "string"
    // For remaining properties, see StoreReadSettings objects
  }

For AzureBlobFSSource, use:

  type = "AzureBlobFSSource"

For AzureDatabricksDeltaLakeSource, use:

  type = "AzureDatabricksDeltaLakeSource"
  exportSettings = {
    type = "string"
  }

For AzureDataExplorerSource, use:

  type = "AzureDataExplorerSource"

For AzureDataLakeStoreSource, use:

  type = "AzureDataLakeStoreSource"

For AzureMariaDBSource, use:

  type = "AzureMariaDBSource"

For AzureMySqlSource, use:

  type = "AzureMySqlSource"

For AzurePostgreSqlSource, use:

  type = "AzurePostgreSqlSource"

For AzureSqlSource, use:

  type = "AzureSqlSource"
  partitionSettings = {}

For AzureTableSource, use:

  type = "AzureTableSource"

For BinarySource, use:

  type = "BinarySource"
  formatSettings = {
    compressionProperties = {
      type = "string"
      // For remaining properties, see CompressionReadSettings objects
    }
    type = "string"
  }
  storeSettings = {
    type = "string"
    // For remaining properties, see StoreReadSettings objects
  }

For BlobSource, use:

  type = "BlobSource"

For CassandraSource, use:

  type = "CassandraSource"
  consistencyLevel = "string"

For CommonDataServiceForAppsSource, use:

  type = "CommonDataServiceForAppsSource"

For ConcurSource, use:

  type = "ConcurSource"

For CosmosDbMongoDbApiSource, use:

  type = "CosmosDbMongoDbApiSource"
  cursorMethods = {}

For CosmosDbSqlApiSource, use:

  type = "CosmosDbSqlApiSource"

For CouchbaseSource, use:

  type = "CouchbaseSource"

For Db2Source, use:

  type = "Db2Source"

For DelimitedTextSource, use:

  type = "DelimitedTextSource"
  formatSettings = {
    compressionProperties = {
      type = "string"
      // For remaining properties, see CompressionReadSettings objects
    }
    type = "string"
  }
  storeSettings = {
    type = "string"
    // For remaining properties, see StoreReadSettings objects
  }

For DocumentDbCollectionSource, use:

  type = "DocumentDbCollectionSource"

For DrillSource, use:

  type = "DrillSource"

For DynamicsAXSource, use:

  type = "DynamicsAXSource"

For DynamicsCrmSource, use:

  type = "DynamicsCrmSource"

For DynamicsSource, use:

  type = "DynamicsSource"

For EloquaSource, use:

  type = "EloquaSource"

For ExcelSource, use:

  type = "ExcelSource"
  storeSettings = {
    type = "string"
    // For remaining properties, see StoreReadSettings objects
  }

For FileSystemSource, use:

  type = "FileSystemSource"

For GoogleAdWordsSource, use:

  type = "GoogleAdWordsSource"

For GoogleBigQuerySource, use:

  type = "GoogleBigQuerySource"

For GoogleBigQueryV2Source, use:

  type = "GoogleBigQueryV2Source"

For GreenplumSource, use:

  type = "GreenplumSource"

For HBaseSource, use:

  type = "HBaseSource"

For HdfsSource, use:

  type = "HdfsSource"
  distcpSettings = {}

For HiveSource, use:

  type = "HiveSource"

For HttpSource, use:

  type = "HttpSource"

For HubspotSource, use:

  type = "HubspotSource"

For ImpalaSource, use:

  type = "ImpalaSource"

For InformixSource, use:

  type = "InformixSource"

For JiraSource, use:

  type = "JiraSource"

For JsonSource, use:

  type = "JsonSource"
  formatSettings = {
    compressionProperties = {
      type = "string"
      // For remaining properties, see CompressionReadSettings objects
    }
    type = "string"
  }
  storeSettings = {
    type = "string"
    // For remaining properties, see StoreReadSettings objects
  }

For LakeHouseTableSource, use:

  type = "LakeHouseTableSource"

For MagentoSource, use:

  type = "MagentoSource"

For MariaDBSource, use:

  type = "MariaDBSource"

For MarketoSource, use:

  type = "MarketoSource"

For MicrosoftAccessSource, use:

  type = "MicrosoftAccessSource"

For MongoDbAtlasSource, use:

  type = "MongoDbAtlasSource"
  cursorMethods = {}

For MongoDbSource, use:

  type = "MongoDbSource"

For MongoDbV2Source, use:

  type = "MongoDbV2Source"
  cursorMethods = {}

For MySqlSource, use:

  type = "MySqlSource"

For NetezzaSource, use:

  type = "NetezzaSource"
  partitionSettings = {}

For ODataSource, use:

  type = "ODataSource"

For OdbcSource, use:

  type = "OdbcSource"

For Office365Source, use:

  type = "Office365Source"

For OracleServiceCloudSource, use:

  type = "OracleServiceCloudSource"

For OracleSource, use:

  type = "OracleSource"
  partitionSettings = {}

For OrcSource, use:

  type = "OrcSource"
  storeSettings = {
    type = "string"
    // For remaining properties, see StoreReadSettings objects
  }

For ParquetSource, use:

  type = "ParquetSource"
  formatSettings = {
    compressionProperties = {
      type = "string"
      // For remaining properties, see CompressionReadSettings objects
    }
    type = "string"
  }
  storeSettings = {
    type = "string"
    // For remaining properties, see StoreReadSettings objects
  }

For PaypalSource, use:

  type = "PaypalSource"

For PhoenixSource, use:

  type = "PhoenixSource"

For PostgreSqlSource, use:

  type = "PostgreSqlSource"

For PostgreSqlV2Source, use:

  type = "PostgreSqlV2Source"

For PrestoSource, use:

  type = "PrestoSource"

For QuickBooksSource, use:

  type = "QuickBooksSource"

For RelationalSource, use:

  type = "RelationalSource"

For ResponsysSource, use:

  type = "ResponsysSource"

For RestSource, use:

  type = "RestSource"

For SalesforceMarketingCloudSource, use:

  type = "SalesforceMarketingCloudSource"

For SalesforceServiceCloudSource, use:

  type = "SalesforceServiceCloudSource"

For SalesforceServiceCloudV2Source, use:

  type = "SalesforceServiceCloudV2Source"

For SalesforceSource, use:

  type = "SalesforceSource"

For SalesforceV2Source, use:

  type = "SalesforceV2Source"

For SapBwSource, use:

  type = "SapBwSource"

For SapCloudForCustomerSource, use:

  type = "SapCloudForCustomerSource"

For SapEccSource, use:

  type = "SapEccSource"

For SapHanaSource, use:

  type = "SapHanaSource"
  partitionSettings = {}

For SapOdpSource, use:

  type = "SapOdpSource"

For SapOpenHubSource, use:

  type = "SapOpenHubSource"

For SapTableSource, use:

  type = "SapTableSource"
  partitionSettings = {}

For ServiceNowSource, use:

  type = "ServiceNowSource"

For ServiceNowV2Source, use:

  type = "ServiceNowV2Source"
  expression = {
    operands = [
      {
      operator = "string"
      type = "string"
      value = "string"
  }

For SharePointOnlineListSource, use:

  type = "SharePointOnlineListSource"

For ShopifySource, use:

  type = "ShopifySource"

For SnowflakeSource, use:

  type = "SnowflakeSource"
  exportSettings = {
    additionalCopyOptions = {}
    additionalFormatOptions = {}
    type = "string"
  }

For SnowflakeV2Source, use:

  type = "SnowflakeV2Source"
  exportSettings = {
    additionalCopyOptions = {}
    additionalFormatOptions = {}
    type = "string"
  }

For SparkSource, use:

  type = "SparkSource"

For SqlDWSource, use:

  type = "SqlDWSource"
  partitionSettings = {}

For SqlMISource, use:

  type = "SqlMISource"
  partitionSettings = {}

For SqlServerSource, use:

  type = "SqlServerSource"
  partitionSettings = {}

For SqlSource, use:

  type = "SqlSource"
  partitionSettings = {}

For SquareSource, use:

  type = "SquareSource"

For SybaseSource, use:

  type = "SybaseSource"

For TeradataSource, use:

  type = "TeradataSource"
  partitionSettings = {}

For VerticaSource, use:

  type = "VerticaSource"

For WarehouseSource, use:

  type = "WarehouseSource"
  partitionSettings = {}

For WebSource, use:

  type = "WebSource"

For XeroSource, use:

  type = "XeroSource"

For XmlSource, use:

  type = "XmlSource"
  formatSettings = {
    compressionProperties = {
      type = "string"
      // For remaining properties, see CompressionReadSettings objects
    }
    type = "string"
  }
  storeSettings = {
    type = "string"
    // For remaining properties, see StoreReadSettings objects
  }

For ZohoSource, use:

  type = "ZohoSource"

StoreReadSettings objects

Set the type property to specify the type of object.

For AmazonS3CompatibleReadSettings, use:

  type = "AmazonS3CompatibleReadSettings"

For AmazonS3ReadSettings, use:

  type = "AmazonS3ReadSettings"

For AzureBlobFSReadSettings, use:

  type = "AzureBlobFSReadSettings"

For AzureBlobStorageReadSettings, use:

  type = "AzureBlobStorageReadSettings"

For AzureDataLakeStoreReadSettings, use:

  type = "AzureDataLakeStoreReadSettings"

For AzureFileStorageReadSettings, use:

  type = "AzureFileStorageReadSettings"

For FileServerReadSettings, use:

  type = "FileServerReadSettings"

For FtpReadSettings, use:

  type = "FtpReadSettings"

For GoogleCloudStorageReadSettings, use:

  type = "GoogleCloudStorageReadSettings"

For HdfsReadSettings, use:

  type = "HdfsReadSettings"
  distcpSettings = {}

For HttpReadSettings, use:

  type = "HttpReadSettings"

For LakeHouseReadSettings, use:

  type = "LakeHouseReadSettings"

For OracleCloudStorageReadSettings, use:

  type = "OracleCloudStorageReadSettings"

For SftpReadSettings, use:

  type = "SftpReadSettings"

CompressionReadSettings objects

Set the type property to specify the type of object.

For TarGZipReadSettings, use:

  type = "TarGZipReadSettings"

For TarReadSettings, use:

  type = "TarReadSettings"

For ZipDeflateReadSettings, use:

  type = "ZipDeflateReadSettings"

SecretBase objects

Set the type property to specify the type of object.

For AzureKeyVaultSecret, use:

  type = "AzureKeyVaultSecret"
  store = {
    parameters = {}
    referenceName = "string"
    type = "LinkedServiceReference"
  }

For SecureString, use:

  type = "SecureString"
  value = "string"

FormatReadSettings objects

Set the type property to specify the type of object.

For BinaryReadSettings, use:

  type = "BinaryReadSettings"
  compressionProperties = {
    type = "string"
    // For remaining properties, see CompressionReadSettings objects
  }

For DelimitedTextReadSettings, use:

  type = "DelimitedTextReadSettings"
  compressionProperties = {
    type = "string"
    // For remaining properties, see CompressionReadSettings objects
  }

For JsonReadSettings, use:

  type = "JsonReadSettings"
  compressionProperties = {
    type = "string"
    // For remaining properties, see CompressionReadSettings objects
  }

For ParquetReadSettings, use:

  type = "ParquetReadSettings"
  compressionProperties = {
    type = "string"
    // For remaining properties, see CompressionReadSettings objects
  }

For XmlReadSettings, use:

  type = "XmlReadSettings"
  compressionProperties = {
    type = "string"
    // For remaining properties, see CompressionReadSettings objects
  }

Property values

factories/pipelines

Name Description Value
type The resource type "Microsoft.DataFactory/factories/pipelines@2018-06-01"
name The resource name string (required)

Character limit: 1-260

Valid characters:
Can't use:
<>*#.%&:\\+?/ or control characters

Start with alphanumeric.
parent_id The ID of the resource that is the parent for this resource. ID for resource of type: factories
properties Properties of the pipeline. Pipeline (required)

Pipeline

Name Description Value
activities List of activities in pipeline. Activity[]
annotations List of tags that can be used for describing the Pipeline. any[]
concurrency The max number of concurrent runs for the pipeline. int

Constraints:
Min value = 1
description The description of the pipeline. string
folder The folder that this Pipeline is in. If not specified, Pipeline will appear at the root level. PipelineFolder
parameters List of parameters for pipeline. ParameterDefinitionSpecification
policy Pipeline Policy. PipelinePolicy
runDimensions Dimensions emitted by Pipeline. PipelineRunDimensions
variables List of variables for pipeline. VariableDefinitionSpecification

Activity

Name Description Value
dependsOn Activity depends on condition. ActivityDependency[]
description Activity description. string
name Activity name. string (required)
onInactiveMarkAs Status result of the activity when the state is set to Inactive. This is an optional property and if not provided when the activity is inactive, the status will be Succeeded by default. "Failed"
"Skipped"
"Succeeded"
state Activity state. This is an optional property and if not provided, the state will be Active by default. "Active"
"Inactive"
userProperties Activity user properties. UserProperty[]
type Set the object type AppendVariable
AzureDataExplorerCommand
AzureFunctionActivity
AzureMLBatchExecution
AzureMLExecutePipeline
AzureMLUpdateResource
Copy
Custom
DatabricksNotebook
DatabricksSparkJar
DatabricksSparkPython
DataLakeAnalyticsU-SQL
Delete
ExecuteDataFlow
ExecutePipeline
ExecuteSSISPackage
ExecuteWranglingDataflow
Fail
Filter
ForEach
GetMetadata
HDInsightHive
HDInsightMapReduce
HDInsightPig
HDInsightSpark
HDInsightStreaming
IfCondition
Lookup
Script
SetVariable
SparkJob
SqlServerStoredProcedure
Switch
SynapseNotebook
Until
Validation
Wait
WebActivity
WebHook (required)

ActivityDependency

Name Description Value
activity Activity name. string (required)
dependencyConditions Match-Condition for the dependency. String array containing any of:
"Completed"
"Failed"
"Skipped"
"Succeeded" (required)
{customized property}

UserProperty

Name Description Value
name User property name. string (required)
value User property value. Type: string (or Expression with resultType string).

AppendVariableActivity

Name Description Value
type Type of activity. "AppendVariable" (required)
typeProperties Append Variable activity properties. AppendVariableActivityTypeProperties (required)

AppendVariableActivityTypeProperties

Name Description Value
value Value to be appended. Type: could be a static value matching type of the variable item or Expression with resultType matching type of the variable item
variableName Name of the variable whose value needs to be appended to. string

AzureDataExplorerCommandActivity

Name Description Value
type Type of activity. "AzureDataExplorerCommand" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Azure Data Explorer command activity properties. AzureDataExplorerCommandActivityTypeProperties (required)

LinkedServiceReference

Name Description Value
parameters Arguments for LinkedService. ParameterValueSpecification
referenceName Reference LinkedService name. string (required)
type Linked service reference type. "LinkedServiceReference" (required)

ParameterValueSpecification

Name Description Value
{customized property}

ActivityPolicy

Name Description Value
retry Maximum ordinary retry attempts. Default is 0. Type: integer (or Expression with resultType integer), minimum: 0.
retryIntervalInSeconds Interval between each retry attempt (in seconds). The default is 30 sec. int

Constraints:
Min value = 30
Max value = 86400
secureInput When set to true, Input from activity is considered as secure and will not be logged to monitoring. bool
secureOutput When set to true, Output from activity is considered as secure and will not be logged to monitoring. bool
timeout Specifies the timeout for the activity to run. The default timeout is 7 days. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
{customized property}

AzureDataExplorerCommandActivityTypeProperties

Name Description Value
command A control command, according to the Azure Data Explorer command syntax. Type: string (or Expression with resultType string).
commandTimeout Control command timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9]))..)

AzureFunctionActivity

Name Description Value
type Type of activity. "AzureFunctionActivity" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Azure Function activity properties. AzureFunctionActivityTypeProperties (required)

AzureFunctionActivityTypeProperties

Name Description Value
body Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string).
functionName Name of the Function that the Azure Function Activity will call. Type: string (or Expression with resultType string)
headers Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). AzureFunctionActivityTypePropertiesHeaders
method Rest API method for target endpoint. "DELETE"
"GET"
"HEAD"
"OPTIONS"
"POST"
"PUT"
"TRACE" (required)

AzureFunctionActivityTypePropertiesHeaders

Name Description Value
{customized property} string

AzureMLBatchExecutionActivity

Name Description Value
type Type of activity. "AzureMLBatchExecution" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Azure ML Batch Execution activity properties. AzureMLBatchExecutionActivityTypeProperties (required)

AzureMLBatchExecutionActivityTypeProperties

Name Description Value
globalParameters Key,Value pairs to be passed to the Azure ML Batch Execution Service endpoint. Keys must match the names of web service parameters defined in the published Azure ML web service. Values will be passed in the GlobalParameters property of the Azure ML batch execution request. AzureMLBatchExecutionActivityTypePropertiesGlobalPar...
webServiceInputs Key,Value pairs, mapping the names of Azure ML endpoint's Web Service Inputs to AzureMLWebServiceFile objects specifying the input Blob locations.. This information will be passed in the WebServiceInputs property of the Azure ML batch execution request. AzureMLBatchExecutionActivityTypePropertiesWebServic...
webServiceOutputs Key,Value pairs, mapping the names of Azure ML endpoint's Web Service Outputs to AzureMLWebServiceFile objects specifying the output Blob locations. This information will be passed in the WebServiceOutputs property of the Azure ML batch execution request. AzureMLBatchExecutionActivityTypePropertiesWebServic...

AzureMLBatchExecutionActivityTypePropertiesGlobalPar...

Name Description Value
{customized property}

AzureMLBatchExecutionActivityTypePropertiesWebServic...

Name Description Value
{customized property} AzureMLWebServiceFile

AzureMLWebServiceFile

Name Description Value
filePath The relative file path, including container name, in the Azure Blob Storage specified by the LinkedService. Type: string (or Expression with resultType string).
linkedServiceName Reference to an Azure Storage LinkedService, where Azure ML WebService Input/Output file located. LinkedServiceReference (required)

AzureMLBatchExecutionActivityTypePropertiesWebServic...

Name Description Value
{customized property} AzureMLWebServiceFile

AzureMLExecutePipelineActivity

Name Description Value
type Type of activity. "AzureMLExecutePipeline" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Azure ML Execute Pipeline activity properties. AzureMLExecutePipelineActivityTypeProperties (required)

AzureMLExecutePipelineActivityTypeProperties

Name Description Value
continueOnStepFailure Whether to continue execution of other steps in the PipelineRun if a step fails. This information will be passed in the continueOnStepFailure property of the published pipeline execution request. Type: boolean (or Expression with resultType boolean).
dataPathAssignments Dictionary used for changing data path assignments without retraining. Values will be passed in the dataPathAssignments property of the published pipeline execution request. Type: object (or Expression with resultType object).
experimentName Run history experiment name of the pipeline run. This information will be passed in the ExperimentName property of the published pipeline execution request. Type: string (or Expression with resultType string).
mlParentRunId The parent Azure ML Service pipeline run id. This information will be passed in the ParentRunId property of the published pipeline execution request. Type: string (or Expression with resultType string).
mlPipelineEndpointId ID of the published Azure ML pipeline endpoint. Type: string (or Expression with resultType string).
mlPipelineId ID of the published Azure ML pipeline. Type: string (or Expression with resultType string).
mlPipelineParameters Key,Value pairs to be passed to the published Azure ML pipeline endpoint. Keys must match the names of pipeline parameters defined in the published pipeline. Values will be passed in the ParameterAssignments property of the published pipeline execution request. Type: object with key value pairs (or Expression with resultType object).
version Version of the published Azure ML pipeline endpoint. Type: string (or Expression with resultType string).

AzureMLUpdateResourceActivity

Name Description Value
type Type of activity. "AzureMLUpdateResource" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Azure ML Update Resource management activity properties. AzureMLUpdateResourceActivityTypeProperties (required)

AzureMLUpdateResourceActivityTypeProperties

Name Description Value
trainedModelFilePath The relative file path in trainedModelLinkedService to represent the .ilearner file that will be uploaded by the update operation. Type: string (or Expression with resultType string).
trainedModelLinkedServiceName Name of Azure Storage linked service holding the .ilearner file that will be uploaded by the update operation. LinkedServiceReference (required)
trainedModelName Name of the Trained Model module in the Web Service experiment to be updated. Type: string (or Expression with resultType string).

CopyActivity

Name Description Value
type Type of activity. "Copy" (required)
inputs List of inputs for the activity. DatasetReference[]
linkedServiceName Linked service reference. LinkedServiceReference
outputs List of outputs for the activity. DatasetReference[]
policy Activity policy. ActivityPolicy
typeProperties Copy activity properties. CopyActivityTypeProperties (required)

DatasetReference

Name Description Value
parameters Arguments for dataset. ParameterValueSpecification
referenceName Reference dataset name. string (required)
type Dataset reference type. "DatasetReference" (required)

CopyActivityTypeProperties

Name Description Value
dataIntegrationUnits Maximum number of data integration units that can be used to perform this data movement. Type: integer (or Expression with resultType integer), minimum: 0.
enableSkipIncompatibleRow Whether to skip incompatible row. Default value is false. Type: boolean (or Expression with resultType boolean).
enableStaging Specifies whether to copy data via an interim staging. Default value is false. Type: boolean (or Expression with resultType boolean).
logSettings Log settings customer needs provide when enabling log. LogSettings
logStorageSettings (Deprecated. Please use LogSettings) Log storage settings customer need to provide when enabling session log. LogStorageSettings
parallelCopies Maximum number of concurrent sessions opened on the source or sink to avoid overloading the data store. Type: integer (or Expression with resultType integer), minimum: 0.
preserve Preserve rules. any[]
preserveRules Preserve Rules. any[]
redirectIncompatibleRowSettings Redirect incompatible row settings when EnableSkipIncompatibleRow is true. RedirectIncompatibleRowSettings
sink Copy activity sink. CopySink (required)
skipErrorFile Specify the fault tolerance for data consistency. SkipErrorFile
source Copy activity source. CopySource (required)
stagingSettings Specifies interim staging settings when EnableStaging is true. StagingSettings
translator Copy activity translator. If not specified, tabular translator is used.
validateDataConsistency Whether to enable Data Consistency validation. Type: boolean (or Expression with resultType boolean).

LogSettings

Name Description Value
copyActivityLogSettings Specifies settings for copy activity log. CopyActivityLogSettings
enableCopyActivityLog Specifies whether to enable copy activity log. Type: boolean (or Expression with resultType boolean).
logLocationSettings Log location settings customer needs to provide when enabling log. LogLocationSettings (required)

CopyActivityLogSettings

Name Description Value
enableReliableLogging Specifies whether to enable reliable logging. Type: boolean (or Expression with resultType boolean).
logLevel Gets or sets the log level, support: Info, Warning. Type: string (or Expression with resultType string).

LogLocationSettings

Name Description Value
linkedServiceName Log storage linked service reference. LinkedServiceReference (required)
path The path to storage for storing detailed logs of activity execution. Type: string (or Expression with resultType string).

LogStorageSettings

Name Description Value
enableReliableLogging Specifies whether to enable reliable logging. Type: boolean (or Expression with resultType boolean).
linkedServiceName Log storage linked service reference. LinkedServiceReference (required)
logLevel Gets or sets the log level, support: Info, Warning. Type: string (or Expression with resultType string).
path The path to storage for storing detailed logs of activity execution. Type: string (or Expression with resultType string).
{customized property}

RedirectIncompatibleRowSettings

Name Description Value
linkedServiceName Name of the Azure Storage, Storage SAS, or Azure Data Lake Store linked service used for redirecting incompatible row. Must be specified if redirectIncompatibleRowSettings is specified. Type: string (or Expression with resultType string).
path The path for storing the redirect incompatible row data. Type: string (or Expression with resultType string).
{customized property}

CopySink

Name Description Value
disableMetricsCollection If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean).
maxConcurrentConnections The maximum concurrent connection count for the sink data store. Type: integer (or Expression with resultType integer).
sinkRetryCount Sink retry count. Type: integer (or Expression with resultType integer).
sinkRetryWait Sink retry wait. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
writeBatchSize Write batch size. Type: integer (or Expression with resultType integer), minimum: 0.
writeBatchTimeout Write batch timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
type Set the object type AvroSink
AzureBlobFSSink
AzureDatabricksDeltaLakeSink
AzureDataExplorerSink
AzureDataLakeStoreSink
AzureMySqlSink
AzurePostgreSqlSink
AzureQueueSink
AzureSearchIndexSink
AzureSqlSink
AzureTableSink
BinarySink
BlobSink
CommonDataServiceForAppsSink
CosmosDbMongoDbApiSink
CosmosDbSqlApiSink
DelimitedTextSink
DocumentDbCollectionSink
DynamicsCrmSink
DynamicsSink
FileSystemSink
InformixSink
JsonSink
LakeHouseTableSink
MicrosoftAccessSink
MongoDbAtlasSink
MongoDbV2Sink
OdbcSink
OracleSink
OrcSink
ParquetSink
RestSink
SalesforceServiceCloudSink
SalesforceServiceCloudV2Sink
SalesforceSink
SalesforceV2Sink
SapCloudForCustomerSink
SnowflakeSink
SnowflakeV2Sink
SqlDWSink
SqlMISink
SqlServerSink
SqlSink
WarehouseSink (required)

AvroSink

Name Description Value
type Copy sink type. "AvroSink" (required)
formatSettings Avro format settings. AvroWriteSettings
storeSettings Avro store settings. StoreWriteSettings

AvroWriteSettings

Name Description Value
fileNamePrefix Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string).
maxRowsPerFile Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer).
recordName Top level record name in write result, which is required in AVRO spec. string
recordNamespace Record namespace in the write result. string
type The write setting type. string (required)
{customized property}

StoreWriteSettings

Name Description Value
copyBehavior The type of copy behavior for copy sink.
disableMetricsCollection If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean).
maxConcurrentConnections The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer).
metadata Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). MetadataItem[]
type Set the object type AzureBlobFSWriteSettings
AzureBlobStorageWriteSettings
AzureDataLakeStoreWriteSettings
AzureFileStorageWriteSettings
FileServerWriteSettings
LakeHouseWriteSettings
SftpWriteSettings (required)

MetadataItem

Name Description Value
name Metadata item key name. Type: string (or Expression with resultType string).
value Metadata item value. Type: string (or Expression with resultType string).

AzureBlobFSWriteSettings

Name Description Value
type The write setting type. "AzureBlobFSWriteSettings" (required)
blockSizeInMB Indicates the block size(MB) when writing data to blob. Type: integer (or Expression with resultType integer).

AzureBlobStorageWriteSettings

Name Description Value
type The write setting type. "AzureBlobStorageWriteSettings" (required)
blockSizeInMB Indicates the block size(MB) when writing data to blob. Type: integer (or Expression with resultType integer).

AzureDataLakeStoreWriteSettings

Name Description Value
type The write setting type. "AzureDataLakeStoreWriteSettings" (required)
expiryDateTime Specifies the expiry time of the written files. The time is applied to the UTC time zone in the format of "2018-12-01T05:00:00Z". Default value is NULL. Type: string (or Expression with resultType string).

AzureFileStorageWriteSettings

Name Description Value
type The write setting type. "AzureFileStorageWriteSettings" (required)

FileServerWriteSettings

Name Description Value
type The write setting type. "FileServerWriteSettings" (required)

LakeHouseWriteSettings

Name Description Value
type The write setting type. "LakeHouseWriteSettings" (required)

SftpWriteSettings

Name Description Value
type The write setting type. "SftpWriteSettings" (required)
operationTimeout Specifies the timeout for writing each chunk to SFTP server. Default value: 01:00:00 (one hour). Type: string (or Expression with resultType string).
useTempFileRename Upload to temporary file(s) and rename. Disable this option if your SFTP server doesn't support rename operation. Type: boolean (or Expression with resultType boolean).

AzureBlobFSSink

Name Description Value
type Copy sink type. "AzureBlobFSSink" (required)
copyBehavior The type of copy behavior for copy sink. Type: string (or Expression with resultType string).
metadata Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). MetadataItem[]

AzureDatabricksDeltaLakeSink

Name Description Value
type Copy sink type. "AzureDatabricksDeltaLakeSink" (required)
importSettings Azure Databricks Delta Lake import settings. AzureDatabricksDeltaLakeImportCommand
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).

AzureDatabricksDeltaLakeImportCommand

Name Description Value
dateFormat Specify the date format for csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string).
timestampFormat Specify the timestamp format for csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string).
type The import setting type. string (required)
{customized property}

AzureDataExplorerSink

Name Description Value
type Copy sink type. "AzureDataExplorerSink" (required)
flushImmediately If set to true, any aggregation will be skipped. Default is false. Type: boolean.
ingestionMappingAsJson An explicit column mapping description provided in a json format. Type: string.
ingestionMappingName A name of a pre-created csv mapping that was defined on the target Kusto table. Type: string.

AzureDataLakeStoreSink

Name Description Value
type Copy sink type. "AzureDataLakeStoreSink" (required)
copyBehavior The type of copy behavior for copy sink. Type: string (or Expression with resultType string).
enableAdlsSingleFileParallel Single File Parallel.

AzureMySqlSink

Name Description Value
type Copy sink type. "AzureMySqlSink" (required)
preCopyScript A query to execute before starting the copy. Type: string (or Expression with resultType string).

AzurePostgreSqlSink

Name Description Value
type Copy sink type. "AzurePostgreSqlSink" (required)
preCopyScript A query to execute before starting the copy. Type: string (or Expression with resultType string).

AzureQueueSink

Name Description Value
type Copy sink type. "AzureQueueSink" (required)

AzureSearchIndexSink

Name Description Value
type Copy sink type. "AzureSearchIndexSink" (required)
writeBehavior Specify the write behavior when upserting documents into Azure Search Index. "Merge"
"Upload"

AzureSqlSink

Name Description Value
type Copy sink type. "AzureSqlSink" (required)
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).
sqlWriterStoredProcedureName SQL writer stored procedure name. Type: string (or Expression with resultType string).
sqlWriterTableType SQL writer table type. Type: string (or Expression with resultType string).
sqlWriterUseTableLock Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean).
storedProcedureParameters SQL stored procedure parameters.
storedProcedureTableTypeParameterName The stored procedure parameter name of the table type. Type: string (or Expression with resultType string).
tableOption The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string).
upsertSettings SQL upsert settings. SqlUpsertSettings
writeBehavior Write behavior when copying data into Azure SQL. Type: SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum)

SqlUpsertSettings

Name Description Value
interimSchemaName Schema name for interim table. Type: string (or Expression with resultType string).
keys Key column names for unique row identification. Type: array of strings (or Expression with resultType array of strings).
useTempDB Specifies whether to use temp db for upsert interim table. Type: boolean (or Expression with resultType boolean).

AzureTableSink

Name Description Value
type Copy sink type. "AzureTableSink" (required)
azureTableDefaultPartitionKeyValue Azure Table default partition key value. Type: string (or Expression with resultType string).
azureTableInsertType Azure Table insert type. Type: string (or Expression with resultType string).
azureTablePartitionKeyName Azure Table partition key name. Type: string (or Expression with resultType string).
azureTableRowKeyName Azure Table row key name. Type: string (or Expression with resultType string).

BinarySink

Name Description Value
type Copy sink type. "BinarySink" (required)
storeSettings Binary store settings. StoreWriteSettings

BlobSink

Name Description Value
type Copy sink type. "BlobSink" (required)
blobWriterAddHeader Blob writer add header. Type: boolean (or Expression with resultType boolean).
blobWriterDateTimeFormat Blob writer date time format. Type: string (or Expression with resultType string).
blobWriterOverwriteFiles Blob writer overwrite files. Type: boolean (or Expression with resultType boolean).
copyBehavior The type of copy behavior for copy sink.
metadata Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). MetadataItem[]

CommonDataServiceForAppsSink

Name Description Value
type Copy sink type. "CommonDataServiceForAppsSink" (required)
alternateKeyName The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string).
ignoreNullValues The flag indicating whether to ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean).
writeBehavior The write behavior for the operation. "Upsert" (required)

CosmosDbMongoDbApiSink

Name Description Value
type Copy sink type. "CosmosDbMongoDbApiSink" (required)
writeBehavior Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string).

CosmosDbSqlApiSink

Name Description Value
type Copy sink type. "CosmosDbSqlApiSink" (required)
writeBehavior Describes how to write data to Azure Cosmos DB. Type: string (or Expression with resultType string). Allowed values: insert and upsert.

DelimitedTextSink

Name Description Value
type Copy sink type. "DelimitedTextSink" (required)
formatSettings DelimitedText format settings. DelimitedTextWriteSettings
storeSettings DelimitedText store settings. StoreWriteSettings

DelimitedTextWriteSettings

Name Description Value
fileExtension The file extension used to create the files. Type: string (or Expression with resultType string).
fileNamePrefix Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string).
maxRowsPerFile Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer).
quoteAllText Indicates whether string values should always be enclosed with quotes. Type: boolean (or Expression with resultType boolean).
type The write setting type. string (required)
{customized property}

DocumentDbCollectionSink

Name Description Value
type Copy sink type. "DocumentDbCollectionSink" (required)
nestingSeparator Nested properties separator. Default is . (dot). Type: string (or Expression with resultType string).
writeBehavior Describes how to write data to Azure Cosmos DB. Type: string (or Expression with resultType string). Allowed values: insert and upsert.

DynamicsCrmSink

Name Description Value
type Copy sink type. "DynamicsCrmSink" (required)
alternateKeyName The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string).
ignoreNullValues The flag indicating whether to ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean).
writeBehavior The write behavior for the operation. "Upsert" (required)

DynamicsSink

Name Description Value
type Copy sink type. "DynamicsSink" (required)
alternateKeyName The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string).
ignoreNullValues The flag indicating whether ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean).
writeBehavior The write behavior for the operation. "Upsert" (required)

FileSystemSink

Name Description Value
type Copy sink type. "FileSystemSink" (required)
copyBehavior The type of copy behavior for copy sink.

InformixSink

Name Description Value
type Copy sink type. "InformixSink" (required)
preCopyScript A query to execute before starting the copy. Type: string (or Expression with resultType string).

JsonSink

Name Description Value
type Copy sink type. "JsonSink" (required)
formatSettings Json format settings. JsonWriteSettings
storeSettings Json store settings. StoreWriteSettings

JsonWriteSettings

Name Description Value
filePattern File pattern of JSON. This setting controls the way a collection of JSON objects will be treated. The default value is 'setOfObjects'. It is case-sensitive.
type The write setting type. string (required)
{customized property}

LakeHouseTableSink

Name Description Value
type Copy sink type. "LakeHouseTableSink" (required)
partitionNameList Specify the partition column names from sink columns. Type: array of objects (or Expression with resultType array of objects).
partitionOption Create partitions in folder structure based on one or multiple columns. Each distinct column value (pair) will be a new partition. Possible values include: "None", "PartitionByKey".
tableActionOption The type of table action for LakeHouse Table sink. Possible values include: "None", "Append", "Overwrite".

MicrosoftAccessSink

Name Description Value
type Copy sink type. "MicrosoftAccessSink" (required)
preCopyScript A query to execute before starting the copy. Type: string (or Expression with resultType string).

MongoDbAtlasSink

Name Description Value
type Copy sink type. "MongoDbAtlasSink" (required)
writeBehavior Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string).

MongoDbV2Sink

Name Description Value
type Copy sink type. "MongoDbV2Sink" (required)
writeBehavior Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string).

OdbcSink

Name Description Value
type Copy sink type. "OdbcSink" (required)
preCopyScript A query to execute before starting the copy. Type: string (or Expression with resultType string).

OracleSink

Name Description Value
type Copy sink type. "OracleSink" (required)
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).

OrcSink

Name Description Value
type Copy sink type. "OrcSink" (required)
formatSettings ORC format settings. OrcWriteSettings
storeSettings ORC store settings. StoreWriteSettings

OrcWriteSettings

Name Description Value
fileNamePrefix Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string).
maxRowsPerFile Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer).
type The write setting type. string (required)
{customized property}

ParquetSink

Name Description Value
type Copy sink type. "ParquetSink" (required)
formatSettings Parquet format settings. ParquetWriteSettings
storeSettings Parquet store settings. StoreWriteSettings

ParquetWriteSettings

Name Description Value
fileNamePrefix Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string).
maxRowsPerFile Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer).
type The write setting type. string (required)
{customized property}

RestSink

Name Description Value
type Copy sink type. "RestSink" (required)
additionalHeaders The additional HTTP headers in the request to the RESTful API. Type: key value pairs (value should be string type).
httpCompressionType Http Compression Type to Send data in compressed format with Optimal Compression Level, Default is None. And The Only Supported option is Gzip. Type: string (or Expression with resultType string).
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
requestInterval The time to await before sending next request, in milliseconds
requestMethod The HTTP method used to call the RESTful API. The default is POST. Type: string (or Expression with resultType string).

SalesforceServiceCloudSink

Name Description Value
type Copy sink type. "SalesforceServiceCloudSink" (required)
externalIdFieldName The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string).
ignoreNullValues The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean).
writeBehavior The write behavior for the operation. Default is Insert. "Insert"
"Upsert"

SalesforceServiceCloudV2Sink

Name Description Value
type Copy sink type. "SalesforceServiceCloudV2Sink" (required)
externalIdFieldName The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string).
ignoreNullValues The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean).
writeBehavior The write behavior for the operation. Default is Insert. "Insert"
"Upsert"

SalesforceSink

Name Description Value
type Copy sink type. "SalesforceSink" (required)
externalIdFieldName The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string).
ignoreNullValues The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean).
writeBehavior The write behavior for the operation. Default is Insert. "Insert"
"Upsert"

SalesforceV2Sink

Name Description Value
type Copy sink type. "SalesforceV2Sink" (required)
externalIdFieldName The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string).
ignoreNullValues The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean).
writeBehavior The write behavior for the operation. Default is Insert. "Insert"
"Upsert"

SapCloudForCustomerSink

Name Description Value
type Copy sink type. "SapCloudForCustomerSink" (required)
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
writeBehavior The write behavior for the operation. Default is 'Insert'. "Insert"
"Update"

SnowflakeSink

Name Description Value
type Copy sink type. "SnowflakeSink" (required)
importSettings Snowflake import settings. SnowflakeImportCopyCommand
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).

SnowflakeImportCopyCommand

Name Description Value
additionalCopyOptions Additional copy options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": "'HH24:MI:SS.FF'" } SnowflakeImportCopyCommandAdditionalCopyOptions
additionalFormatOptions Additional format options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalFormatOptions": { "FORCE": "TRUE", "LOAD_UNCERTAIN_FILES": "'FALSE'" } SnowflakeImportCopyCommandAdditionalFormatOptions
type The import setting type. string (required)
{customized property}

SnowflakeImportCopyCommandAdditionalCopyOptions

Name Description Value
{customized property}

SnowflakeImportCopyCommandAdditionalFormatOptions

Name Description Value
{customized property}

SnowflakeV2Sink

Name Description Value
type Copy sink type. "SnowflakeV2Sink" (required)
importSettings Snowflake import settings. SnowflakeImportCopyCommand
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).

SqlDWSink

Name Description Value
type Copy sink type. "SqlDWSink" (required)
allowCopyCommand Indicates to use Copy Command to copy data into SQL Data Warehouse. Type: boolean (or Expression with resultType boolean).
allowPolyBase Indicates to use PolyBase to copy data into SQL Data Warehouse when applicable. Type: boolean (or Expression with resultType boolean).
copyCommandSettings Specifies Copy Command related settings when allowCopyCommand is true. DWCopyCommandSettings
polyBaseSettings Specifies PolyBase-related settings when allowPolyBase is true. PolybaseSettings
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).
sqlWriterUseTableLock Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean).
tableOption The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string).
upsertSettings SQL DW upsert settings. SqlDWUpsertSettings
writeBehavior Write behavior when copying data into azure SQL DW. Type: SqlDWWriteBehaviorEnum (or Expression with resultType SqlDWWriteBehaviorEnum)

DWCopyCommandSettings

Name Description Value
additionalOptions Additional options directly passed to SQL DW in Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalOptions": { "MAXERRORS": "1000", "DATEFORMAT": "'ymd'" } DWCopyCommandSettingsAdditionalOptions
defaultValues Specifies the default values for each target column in SQL DW. The default values in the property overwrite the DEFAULT constraint set in the DB, and identity column cannot have a default value. Type: array of objects (or Expression with resultType array of objects). DWCopyCommandDefaultValue[]

DWCopyCommandSettingsAdditionalOptions

Name Description Value
{customized property} string

DWCopyCommandDefaultValue

Name Description Value
columnName Column name. Type: object (or Expression with resultType string).
defaultValue The default value of the column. Type: object (or Expression with resultType string).

PolybaseSettings

Name Description Value
rejectSampleValue Determines the number of rows to attempt to retrieve before the PolyBase recalculates the percentage of rejected rows. Type: integer (or Expression with resultType integer), minimum: 0.
rejectType Reject type. "percentage"
"value"
rejectValue Specifies the value or the percentage of rows that can be rejected before the query fails. Type: number (or Expression with resultType number), minimum: 0.
useTypeDefault Specifies how to handle missing values in delimited text files when PolyBase retrieves data from the text file. Type: boolean (or Expression with resultType boolean).
{customized property}

SqlDWUpsertSettings

Name Description Value
interimSchemaName Schema name for interim table. Type: string (or Expression with resultType string).
keys Key column names for unique row identification. Type: array of strings (or Expression with resultType array of strings).

SqlMISink

Name Description Value
type Copy sink type. "SqlMISink" (required)
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).
sqlWriterStoredProcedureName SQL writer stored procedure name. Type: string (or Expression with resultType string).
sqlWriterTableType SQL writer table type. Type: string (or Expression with resultType string).
sqlWriterUseTableLock Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean).
storedProcedureParameters SQL stored procedure parameters.
storedProcedureTableTypeParameterName The stored procedure parameter name of the table type. Type: string (or Expression with resultType string).
tableOption The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string).
upsertSettings SQL upsert settings. SqlUpsertSettings
writeBehavior White behavior when copying data into azure SQL MI. Type: string (or Expression with resultType string)

SqlServerSink

Name Description Value
type Copy sink type. "SqlServerSink" (required)
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).
sqlWriterStoredProcedureName SQL writer stored procedure name. Type: string (or Expression with resultType string).
sqlWriterTableType SQL writer table type. Type: string (or Expression with resultType string).
sqlWriterUseTableLock Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean).
storedProcedureParameters SQL stored procedure parameters.
storedProcedureTableTypeParameterName The stored procedure parameter name of the table type. Type: string (or Expression with resultType string).
tableOption The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string).
upsertSettings SQL upsert settings. SqlUpsertSettings
writeBehavior Write behavior when copying data into sql server. Type: string (or Expression with resultType string).

SqlSink

Name Description Value
type Copy sink type. "SqlSink" (required)
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).
sqlWriterStoredProcedureName SQL writer stored procedure name. Type: string (or Expression with resultType string).
sqlWriterTableType SQL writer table type. Type: string (or Expression with resultType string).
sqlWriterUseTableLock Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean).
storedProcedureParameters SQL stored procedure parameters.
storedProcedureTableTypeParameterName The stored procedure parameter name of the table type. Type: string (or Expression with resultType string).
tableOption The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string).
upsertSettings SQL upsert settings. SqlUpsertSettings
writeBehavior Write behavior when copying data into sql. Type: string (or Expression with resultType string).

WarehouseSink

Name Description Value
type Copy sink type. "WarehouseSink" (required)
allowCopyCommand Indicates to use Copy Command to copy data into SQL Data Warehouse. Type: boolean (or Expression with resultType boolean).
copyCommandSettings Specifies Copy Command related settings when allowCopyCommand is true. DWCopyCommandSettings
preCopyScript SQL pre-copy script. Type: string (or Expression with resultType string).
tableOption The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string).
writeBehavior Write behavior when copying data into azure Microsoft Fabric Data Warehouse. Type: DWWriteBehaviorEnum (or Expression with resultType DWWriteBehaviorEnum)

SkipErrorFile

Name Description Value
dataInconsistency Skip if source/sink file changed by other concurrent write. Default is false. Type: boolean (or Expression with resultType boolean).
fileMissing Skip if file is deleted by other client during copy. Default is true. Type: boolean (or Expression with resultType boolean).

CopySource

Name Description Value
disableMetricsCollection If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean).
maxConcurrentConnections The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer).
sourceRetryCount Source retry count. Type: integer (or Expression with resultType integer).
sourceRetryWait Source retry wait. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
type Set the object type AmazonMWSSource
AmazonRdsForOracleSource
AmazonRdsForSqlServerSource
AmazonRedshiftSource
AvroSource
AzureBlobFSSource
AzureDatabricksDeltaLakeSource
AzureDataExplorerSource
AzureDataLakeStoreSource
AzureMariaDBSource
AzureMySqlSource
AzurePostgreSqlSource
AzureSqlSource
AzureTableSource
BinarySource
BlobSource
CassandraSource
CommonDataServiceForAppsSource
ConcurSource
CosmosDbMongoDbApiSource
CosmosDbSqlApiSource
CouchbaseSource
Db2Source
DelimitedTextSource
DocumentDbCollectionSource
DrillSource
DynamicsAXSource
DynamicsCrmSource
DynamicsSource
EloquaSource
ExcelSource
FileSystemSource
GoogleAdWordsSource
GoogleBigQuerySource
GoogleBigQueryV2Source
GreenplumSource
HBaseSource
HdfsSource
HiveSource
HttpSource
HubspotSource
ImpalaSource
InformixSource
JiraSource
JsonSource
LakeHouseTableSource
MagentoSource
MariaDBSource
MarketoSource
MicrosoftAccessSource
MongoDbAtlasSource
MongoDbSource
MongoDbV2Source
MySqlSource
NetezzaSource
ODataSource
OdbcSource
Office365Source
OracleServiceCloudSource
OracleSource
OrcSource
ParquetSource
PaypalSource
PhoenixSource
PostgreSqlSource
PostgreSqlV2Source
PrestoSource
QuickBooksSource
RelationalSource
ResponsysSource
RestSource
SalesforceMarketingCloudSource
SalesforceServiceCloudSource
SalesforceServiceCloudV2Source
SalesforceSource
SalesforceV2Source
SapBwSource
SapCloudForCustomerSource
SapEccSource
SapHanaSource
SapOdpSource
SapOpenHubSource
SapTableSource
ServiceNowSource
ServiceNowV2Source
SharePointOnlineListSource
ShopifySource
SnowflakeSource
SnowflakeV2Source
SparkSource
SqlDWSource
SqlMISource
SqlServerSource
SqlSource
SquareSource
SybaseSource
TeradataSource
VerticaSource
WarehouseSource
WebSource
XeroSource
XmlSource
ZohoSource (required)

AmazonMWSSource

Name Description Value
type Copy source type. "AmazonMWSSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

AmazonRdsForOracleSource

Name Description Value
type Copy source type. "AmazonRdsForOracleSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
oracleReaderQuery AmazonRdsForOracle reader query. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for AmazonRdsForOracle read in parallel. Type: string (or Expression with resultType string).
partitionSettings The settings that will be leveraged for AmazonRdsForOracle source partitioning. AmazonRdsForOraclePartitionSettings
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

AmazonRdsForOraclePartitionSettings

Name Description Value
partitionColumnName The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).
partitionLowerBound The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).
partitionNames Names of the physical partitions of AmazonRdsForOracle table.
partitionUpperBound The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).

AmazonRdsForSqlServerSource

Name Description Value
type Copy source type. "AmazonRdsForSqlServerSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
isolationLevel Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange".
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
produceAdditionalTypes Which additional types to produce.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
sqlReaderQuery SQL reader query. Type: string (or Expression with resultType string).
sqlReaderStoredProcedureName Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string).
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}".

SqlPartitionSettings

Name Description Value
partitionColumnName The name of the column in integer or datetime type that will be used for proceeding partitioning. If not specified, the primary key of the table is auto-detected and used as the partition column. Type: string (or Expression with resultType string).
partitionLowerBound The minimum value of the partition column for partition range splitting. This value is used to decide the partition stride, not for filtering the rows in table. All rows in the table or query result will be partitioned and copied. Type: string (or Expression with resultType string).
partitionUpperBound The maximum value of the partition column for partition range splitting. This value is used to decide the partition stride, not for filtering the rows in table. All rows in the table or query result will be partitioned and copied. Type: string (or Expression with resultType string).

AmazonRedshiftSource

Name Description Value
type Copy source type. "AmazonRedshiftSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
redshiftUnloadSettings The Amazon S3 settings needed for the interim Amazon S3 when copying from Amazon Redshift with unload. With this, data from Amazon Redshift source will be unloaded into S3 first and then copied into the targeted sink from the interim S3. RedshiftUnloadSettings

RedshiftUnloadSettings

Name Description Value
bucketName The bucket of the interim Amazon S3 which will be used to store the unloaded data from Amazon Redshift source. The bucket must be in the same region as the Amazon Redshift source. Type: string (or Expression with resultType string).
s3LinkedServiceName The name of the Amazon S3 linked service which will be used for the unload operation when copying from the Amazon Redshift source. LinkedServiceReference (required)

AvroSource

Name Description Value
type Copy source type. "AvroSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
storeSettings Avro store settings. StoreReadSettings

StoreReadSettings

Name Description Value
disableMetricsCollection If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean).
maxConcurrentConnections The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer).
type Set the object type AmazonS3CompatibleReadSettings
AmazonS3ReadSettings
AzureBlobFSReadSettings
AzureBlobStorageReadSettings
AzureDataLakeStoreReadSettings
AzureFileStorageReadSettings
FileServerReadSettings
FtpReadSettings
GoogleCloudStorageReadSettings
HdfsReadSettings
HttpReadSettings
LakeHouseReadSettings
OracleCloudStorageReadSettings
SftpReadSettings (required)

AmazonS3CompatibleReadSettings

Name Description Value
type The read setting type. "AmazonS3CompatibleReadSettings" (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
prefix The prefix filter for the S3 Compatible object name. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName Amazon S3 Compatible wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath Amazon S3 Compatible wildcardFolderPath. Type: string (or Expression with resultType string).

AmazonS3ReadSettings

Name Description Value
type The read setting type. "AmazonS3ReadSettings" (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
prefix The prefix filter for the S3 object name. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName AmazonS3 wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath AmazonS3 wildcardFolderPath. Type: string (or Expression with resultType string).

AzureBlobFSReadSettings

Name Description Value
type The read setting type. "AzureBlobFSReadSettings" (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName Azure blobFS wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath Azure blobFS wildcardFolderPath. Type: string (or Expression with resultType string).

AzureBlobStorageReadSettings

Name Description Value
type The read setting type. "AzureBlobStorageReadSettings" (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
prefix The prefix filter for the Azure Blob name. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName Azure blob wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath Azure blob wildcardFolderPath. Type: string (or Expression with resultType string).

AzureDataLakeStoreReadSettings

Name Description Value
type The read setting type. "AzureDataLakeStoreReadSettings" (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
listAfter Lists files after the value (exclusive) based on file/folder names’ lexicographical order. Applies under the folderPath in data set, and filter files/sub-folders under the folderPath. Type: string (or Expression with resultType string).
listBefore Lists files before the value (inclusive) based on file/folder names’ lexicographical order. Applies under the folderPath in data set, and filter files/sub-folders under the folderPath. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName ADLS wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath ADLS wildcardFolderPath. Type: string (or Expression with resultType string).

AzureFileStorageReadSettings

Name Description Value
type The read setting type. "AzureFileStorageReadSettings" (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
prefix The prefix filter for the Azure File name starting from root path. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName Azure File Storage wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath Azure File Storage wildcardFolderPath. Type: string (or Expression with resultType string).

FileServerReadSettings

Name Description Value
type The read setting type. "FileServerReadSettings" (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileFilter Specify a filter to be used to select a subset of files in the folderPath rather than all files. Type: string (or Expression with resultType string).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName FileServer wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath FileServer wildcardFolderPath. Type: string (or Expression with resultType string).

FtpReadSettings

Name Description Value
type The read setting type. "FtpReadSettings" (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
disableChunking If true, disable parallel reading within each file. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
useBinaryTransfer Specify whether to use binary transfer mode for FTP stores. Type: boolean (or Expression with resultType boolean).
wildcardFileName Ftp wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath Ftp wildcardFolderPath. Type: string (or Expression with resultType string).

GoogleCloudStorageReadSettings

Name Description Value
type The read setting type. "GoogleCloudStorageReadSettings" (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
prefix The prefix filter for the Google Cloud Storage object name. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName Google Cloud Storage wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath Google Cloud Storage wildcardFolderPath. Type: string (or Expression with resultType string).

HdfsReadSettings

Name Description Value
type The read setting type. "HdfsReadSettings" (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
distcpSettings Specifies Distcp-related settings. DistcpSettings
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName HDFS wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath HDFS wildcardFolderPath. Type: string (or Expression with resultType string).

DistcpSettings

Name Description Value
distcpOptions Specifies the Distcp options. Type: string (or Expression with resultType string).
resourceManagerEndpoint Specifies the Yarn ResourceManager endpoint. Type: string (or Expression with resultType string).
tempScriptPath Specifies an existing folder path which will be used to store temp Distcp command script. The script file is generated by ADF and will be removed after Copy job finished. Type: string (or Expression with resultType string).

HttpReadSettings

Name Description Value
type The read setting type. "HttpReadSettings" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
additionalHeaders The additional HTTP headers in the request to the RESTful API. Type: string (or Expression with resultType string).
requestBody The HTTP request body to the RESTful API if requestMethod is POST. Type: string (or Expression with resultType string).
requestMethod The HTTP method used to call the RESTful API. The default is GET. Type: string (or Expression with resultType string).
requestTimeout Specifies the timeout for a HTTP client to get HTTP response from HTTP server. Type: string (or Expression with resultType string).

LakeHouseReadSettings

Name Description Value
type The read setting type. "LakeHouseReadSettings" (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName Microsoft Fabric LakeHouse Files wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath Microsoft Fabric LakeHouse Files wildcardFolderPath. Type: string (or Expression with resultType string).

OracleCloudStorageReadSettings

Name Description Value
type The read setting type. "OracleCloudStorageReadSettings" (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
prefix The prefix filter for the Oracle Cloud Storage object name. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName Oracle Cloud Storage wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath Oracle Cloud Storage wildcardFolderPath. Type: string (or Expression with resultType string).

SftpReadSettings

Name Description Value
type The read setting type. "SftpReadSettings" (required)
deleteFilesAfterCompletion Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean).
disableChunking If true, disable parallel reading within each file. Default is false. Type: boolean (or Expression with resultType boolean).
enablePartitionDiscovery Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean).
fileListPath Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string).
modifiedDatetimeEnd The end of file's modified datetime. Type: string (or Expression with resultType string).
modifiedDatetimeStart The start of file's modified datetime. Type: string (or Expression with resultType string).
partitionRootPath Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
wildcardFileName Sftp wildcardFileName. Type: string (or Expression with resultType string).
wildcardFolderPath Sftp wildcardFolderPath. Type: string (or Expression with resultType string).

AzureBlobFSSource

Name Description Value
type Copy source type. "AzureBlobFSSource" (required)
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
skipHeaderLineCount Number of header lines to skip from each blob. Type: integer (or Expression with resultType integer).
treatEmptyAsNull Treat empty as null. Type: boolean (or Expression with resultType boolean).

AzureDatabricksDeltaLakeSource

Name Description Value
type Copy source type. "AzureDatabricksDeltaLakeSource" (required)
exportSettings Azure Databricks Delta Lake export settings. AzureDatabricksDeltaLakeExportCommand
query Azure Databricks Delta Lake Sql query. Type: string (or Expression with resultType string).

AzureDatabricksDeltaLakeExportCommand

Name Description Value
dateFormat Specify the date format for the csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string).
timestampFormat Specify the timestamp format for the csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string).
type The export setting type. string (required)
{customized property}

AzureDataExplorerSource

Name Description Value
type Copy source type. "AzureDataExplorerSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
noTruncation The name of the Boolean option that controls whether truncation is applied to result-sets that go beyond a certain row-count limit.
query Database query. Should be a Kusto Query Language (KQL) query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9]))..

AzureDataLakeStoreSource

Name Description Value
type Copy source type. "AzureDataLakeStoreSource" (required)
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).

AzureMariaDBSource

Name Description Value
type Copy source type. "AzureMariaDBSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

AzureMySqlSource

Name Description Value
type Copy source type. "AzureMySqlSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

AzurePostgreSqlSource

Name Description Value
type Copy source type. "AzurePostgreSqlSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

AzureSqlSource

Name Description Value
type Copy source type. "AzureSqlSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
isolationLevel Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string).
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
produceAdditionalTypes Which additional types to produce.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
sqlReaderQuery SQL reader query. Type: string (or Expression with resultType string).
sqlReaderStoredProcedureName Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string).
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}".

AzureTableSource

Name Description Value
type Copy source type. "AzureTableSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
azureTableSourceIgnoreTableNotFound Azure Table source ignore table not found. Type: boolean (or Expression with resultType boolean).
azureTableSourceQuery Azure Table source query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

BinarySource

Name Description Value
type Copy source type. "BinarySource" (required)
formatSettings Binary format settings. BinaryReadSettings
storeSettings Binary store settings. StoreReadSettings

BinaryReadSettings

Name Description Value
compressionProperties Compression settings. CompressionReadSettings
type The read setting type. string (required)
{customized property}

CompressionReadSettings

Name Description Value
type Set the object type TarGZipReadSettings
TarReadSettings
ZipDeflateReadSettings (required)

TarGZipReadSettings

Name Description Value
type The Compression setting type. "TarGZipReadSettings" (required)
preserveCompressionFileNameAsFolder Preserve the compression file name as folder path. Type: boolean (or Expression with resultType boolean).

TarReadSettings

Name Description Value
type The Compression setting type. "TarReadSettings" (required)
preserveCompressionFileNameAsFolder Preserve the compression file name as folder path. Type: boolean (or Expression with resultType boolean).

ZipDeflateReadSettings

Name Description Value
type The Compression setting type. "ZipDeflateReadSettings" (required)
preserveZipFileNameAsFolder Preserve the zip file name as folder path. Type: boolean (or Expression with resultType boolean).

BlobSource

Name Description Value
type Copy source type. "BlobSource" (required)
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).
skipHeaderLineCount Number of header lines to skip from each blob. Type: integer (or Expression with resultType integer).
treatEmptyAsNull Treat empty as null. Type: boolean (or Expression with resultType boolean).

CassandraSource

Name Description Value
type Copy source type. "CassandraSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
consistencyLevel The consistency level specifies how many Cassandra servers must respond to a read request before returning data to the client application. Cassandra checks the specified number of Cassandra servers for data to satisfy the read request. Must be one of cassandraSourceReadConsistencyLevels. The default value is 'ONE'. It is case-insensitive. "ALL"
"EACH_QUORUM"
"LOCAL_ONE"
"LOCAL_QUORUM"
"LOCAL_SERIAL"
"ONE"
"QUORUM"
"SERIAL"
"THREE"
"TWO"
query Database query. Should be a SQL-92 query expression or Cassandra Query Language (CQL) command. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

CommonDataServiceForAppsSource

Name Description Value
type Copy source type. "CommonDataServiceForAppsSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query FetchXML is a proprietary query language that is used in Microsoft Common Data Service for Apps (online & on-premises). Type: string (or Expression with resultType string).

ConcurSource

Name Description Value
type Copy source type. "ConcurSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

CosmosDbMongoDbApiSource

Name Description Value
type Copy source type. "CosmosDbMongoDbApiSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
batchSize Specifies the number of documents to return in each batch of the response from MongoDB instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer).
cursorMethods Cursor methods for Mongodb query. MongoDbCursorMethodsProperties
filter Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

MongoDbCursorMethodsProperties

Name Description Value
limit Specifies the maximum number of documents the server returns. limit() is analogous to the LIMIT statement in a SQL database. Type: integer (or Expression with resultType integer).
project Specifies the fields to return in the documents that match the query filter. To return all fields in the matching documents, omit this parameter. Type: string (or Expression with resultType string).
skip Specifies the how many documents skipped and where MongoDB begins returning results. This approach may be useful in implementing paginated results. Type: integer (or Expression with resultType integer).
sort Specifies the order in which the query returns matching documents. Type: string (or Expression with resultType string). Type: string (or Expression with resultType string).
{customized property}

CosmosDbSqlApiSource

Name Description Value
type Copy source type. "CosmosDbSqlApiSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
detectDatetime Whether detect primitive values as datetime values. Type: boolean (or Expression with resultType boolean).
pageSize Page size of the result. Type: integer (or Expression with resultType integer).
preferredRegions Preferred regions. Type: array of strings (or Expression with resultType array of strings).
query SQL API query. Type: string (or Expression with resultType string).

CouchbaseSource

Name Description Value
type Copy source type. "CouchbaseSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

Db2Source

Name Description Value
type Copy source type. "Db2Source" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

DelimitedTextSource

Name Description Value
type Copy source type. "DelimitedTextSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
formatSettings DelimitedText format settings. DelimitedTextReadSettings
storeSettings DelimitedText store settings. StoreReadSettings

DelimitedTextReadSettings

Name Description Value
compressionProperties Compression settings. CompressionReadSettings
skipLineCount Indicates the number of non-empty rows to skip when reading data from input files. Type: integer (or Expression with resultType integer).
type The read setting type. string (required)
{customized property}

DocumentDbCollectionSource

Name Description Value
type Copy source type. "DocumentDbCollectionSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
nestingSeparator Nested properties separator. Type: string (or Expression with resultType string).
query Documents query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

DrillSource

Name Description Value
type Copy source type. "DrillSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

DynamicsAXSource

Name Description Value
type Copy source type. "DynamicsAXSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

DynamicsCrmSource

Name Description Value
type Copy source type. "DynamicsCrmSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query FetchXML is a proprietary query language that is used in Microsoft Dynamics CRM (online & on-premises). Type: string (or Expression with resultType string).

DynamicsSource

Name Description Value
type Copy source type. "DynamicsSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query FetchXML is a proprietary query language that is used in Microsoft Dynamics (online & on-premises). Type: string (or Expression with resultType string).

EloquaSource

Name Description Value
type Copy source type. "EloquaSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

ExcelSource

Name Description Value
type Copy source type. "ExcelSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
storeSettings Excel store settings. StoreReadSettings

FileSystemSource

Name Description Value
type Copy source type. "FileSystemSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).

GoogleAdWordsSource

Name Description Value
type Copy source type. "GoogleAdWordsSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

GoogleBigQuerySource

Name Description Value
type Copy source type. "GoogleBigQuerySource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

GoogleBigQueryV2Source

Name Description Value
type Copy source type. "GoogleBigQueryV2Source" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

GreenplumSource

Name Description Value
type Copy source type. "GreenplumSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

HBaseSource

Name Description Value
type Copy source type. "HBaseSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

HdfsSource

Name Description Value
type Copy source type. "HdfsSource" (required)
distcpSettings Specifies Distcp-related settings. DistcpSettings
recursive If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean).

HiveSource

Name Description Value
type Copy source type. "HiveSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

HttpSource

Name Description Value
type Copy source type. "HttpSource" (required)
httpRequestTimeout Specifies the timeout for a HTTP client to get HTTP response from HTTP server. The default value is equivalent to System.Net.HttpWebRequest.Timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

HubspotSource

Name Description Value
type Copy source type. "HubspotSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

ImpalaSource

Name Description Value
type Copy source type. "ImpalaSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

InformixSource

Name Description Value
type Copy source type. "InformixSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

JiraSource

Name Description Value
type Copy source type. "JiraSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

JsonSource

Name Description Value
type Copy source type. "JsonSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
formatSettings Json format settings. JsonReadSettings
storeSettings Json store settings. StoreReadSettings

JsonReadSettings

Name Description Value
compressionProperties Compression settings. CompressionReadSettings
type The read setting type. string (required)
{customized property}

LakeHouseTableSource

Name Description Value
type Copy source type. "LakeHouseTableSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
timestampAsOf Query an older snapshot by timestamp. Type: string (or Expression with resultType string).
versionAsOf Query an older snapshot by version. Type: integer (or Expression with resultType integer).

MagentoSource

Name Description Value
type Copy source type. "MagentoSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

MariaDBSource

Name Description Value
type Copy source type. "MariaDBSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

MarketoSource

Name Description Value
type Copy source type. "MarketoSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

MicrosoftAccessSource

Name Description Value
type Copy source type. "MicrosoftAccessSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).

MongoDbAtlasSource

Name Description Value
type Copy source type. "MongoDbAtlasSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
batchSize Specifies the number of documents to return in each batch of the response from MongoDB Atlas instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer).
cursorMethods Cursor methods for Mongodb query MongoDbCursorMethodsProperties
filter Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

MongoDbSource

Name Description Value
type Copy source type. "MongoDbSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Should be a SQL-92 query expression. Type: string (or Expression with resultType string).

MongoDbV2Source

Name Description Value
type Copy source type. "MongoDbV2Source" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
batchSize Specifies the number of documents to return in each batch of the response from MongoDB instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer).
cursorMethods Cursor methods for Mongodb query MongoDbCursorMethodsProperties
filter Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

MySqlSource

Name Description Value
type Copy source type. "MySqlSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

NetezzaSource

Name Description Value
type Copy source type. "NetezzaSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
partitionOption The partition mechanism that will be used for Netezza read in parallel. Possible values include: "None", "DataSlice", "DynamicRange".
partitionSettings The settings that will be leveraged for Netezza source partitioning. NetezzaPartitionSettings
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

NetezzaPartitionSettings

Name Description Value
partitionColumnName The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).
partitionLowerBound The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).
partitionUpperBound The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).

ODataSource

Name Description Value
type Copy source type. "ODataSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
query OData query. For example, "$top=1". Type: string (or Expression with resultType string).

OdbcSource

Name Description Value
type Copy source type. "OdbcSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

Office365Source

Name Description Value
type Copy source type. "Office365Source" (required)
allowedGroups The groups containing all the users. Type: array of strings (or Expression with resultType array of strings).
dateFilterColumn The Column to apply the {paramref name="StartTime"/} and {paramref name="EndTime"/}. Type: string (or Expression with resultType string).
endTime End time of the requested range for this dataset. Type: string (or Expression with resultType string).
outputColumns The columns to be read out from the Office 365 table. Type: array of objects (or Expression with resultType array of objects). itemType: OutputColumn. Example: [ { "name": "Id" }, { "name": "CreatedDateTime" } ]
startTime Start time of the requested range for this dataset. Type: string (or Expression with resultType string).
userScopeFilterUri The user scope uri. Type: string (or Expression with resultType string).

OracleServiceCloudSource

Name Description Value
type Copy source type. "OracleServiceCloudSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

OracleSource

Name Description Value
type Copy source type. "OracleSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
oracleReaderQuery Oracle reader query. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for Oracle read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange".
partitionSettings The settings that will be leveraged for Oracle source partitioning. OraclePartitionSettings
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

OraclePartitionSettings

Name Description Value
partitionColumnName The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).
partitionLowerBound The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).
partitionNames Names of the physical partitions of Oracle table.
partitionUpperBound The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).

OrcSource

Name Description Value
type Copy source type. "OrcSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
storeSettings ORC store settings. StoreReadSettings

ParquetSource

Name Description Value
type Copy source type. "ParquetSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
formatSettings Parquet format settings. ParquetReadSettings
storeSettings Parquet store settings. StoreReadSettings

ParquetReadSettings

Name Description Value
compressionProperties Compression settings. CompressionReadSettings
type The read setting type. string (required)
{customized property}

PaypalSource

Name Description Value
type Copy source type. "PaypalSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

PhoenixSource

Name Description Value
type Copy source type. "PhoenixSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

PostgreSqlSource

Name Description Value
type Copy source type. "PostgreSqlSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

PostgreSqlV2Source

Name Description Value
type Copy source type. "PostgreSqlV2Source" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

PrestoSource

Name Description Value
type Copy source type. "PrestoSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

QuickBooksSource

Name Description Value
type Copy source type. "QuickBooksSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

RelationalSource

Name Description Value
type Copy source type. "RelationalSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).

ResponsysSource

Name Description Value
type Copy source type. "ResponsysSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

RestSource

Name Description Value
type Copy source type. "RestSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: key value pairs (value should be string type).
additionalHeaders The additional HTTP headers in the request to the RESTful API. Type: string (or Expression with resultType string).
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
paginationRules The pagination rules to compose next page requests. Type: string (or Expression with resultType string).
requestBody The HTTP request body to the RESTful API if requestMethod is POST. Type: string (or Expression with resultType string).
requestInterval The time to await before sending next page request.
requestMethod The HTTP method used to call the RESTful API. The default is GET. Type: string (or Expression with resultType string).

SalesforceMarketingCloudSource

Name Description Value
type Copy source type. "SalesforceMarketingCloudSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

SalesforceServiceCloudSource

Name Description Value
type Copy source type. "SalesforceServiceCloudSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
readBehavior The read behavior for the operation. Default is Query. Allowed values: Query/QueryAll. Type: string (or Expression with resultType string).

SalesforceServiceCloudV2Source

Name Description Value
type Copy source type. "SalesforceServiceCloudV2Source" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
includeDeletedObjects This property control whether query result contains Deleted objects. Default is false. Type: boolean (or Expression with resultType boolean).
SOQLQuery Database query. Type: string (or Expression with resultType string).

SalesforceSource

Name Description Value
type Copy source type. "SalesforceSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
readBehavior The read behavior for the operation. Default is Query. Allowed values: Query/QueryAll. Type: string (or Expression with resultType string).

SalesforceV2Source

Name Description Value
type Copy source type. "SalesforceV2Source" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
includeDeletedObjects This property control whether query result contains Deleted objects. Default is false. Type: boolean (or Expression with resultType boolean).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
SOQLQuery Database query. Type: string (or Expression with resultType string).

SapBwSource

Name Description Value
type Copy source type. "SapBwSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query MDX query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

SapCloudForCustomerSource

Name Description Value
type Copy source type. "SapCloudForCustomerSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
query SAP Cloud for Customer OData query. For example, "$top=1". Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

SapEccSource

Name Description Value
type Copy source type. "SapEccSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
httpRequestTimeout The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
query SAP ECC OData query. For example, "$top=1". Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

SapHanaSource

Name Description Value
type Copy source type. "SapHanaSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
packetSize The packet size of data read from SAP HANA. Type: integer(or Expression with resultType integer).
partitionOption The partition mechanism that will be used for SAP HANA read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "SapHanaDynamicRange".
partitionSettings The settings that will be leveraged for SAP HANA source partitioning. SapHanaPartitionSettings
query SAP HANA Sql query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

SapHanaPartitionSettings

Name Description Value
partitionColumnName The name of the column that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).

SapOdpSource

Name Description Value
type Copy source type. "SapOdpSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
extractionMode The extraction mode. Allowed value include: Full, Delta and Recovery. The default value is Full. Type: string (or Expression with resultType string).
projection Specifies the columns to be selected from source data. Type: array of objects(projection) (or Expression with resultType array of objects).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
selection Specifies the selection conditions from source data. Type: array of objects(selection) (or Expression with resultType array of objects).
subscriberProcess The subscriber process to manage the delta process. Type: string (or Expression with resultType string).

SapOpenHubSource

Name Description Value
type Copy source type. "SapOpenHubSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
baseRequestId The ID of request for delta loading. Once it is set, only data with requestId larger than the value of this property will be retrieved. The default value is 0. Type: integer (or Expression with resultType integer ).
customRfcReadTableFunctionModule Specifies the custom RFC function module that will be used to read data from SAP Table. Type: string (or Expression with resultType string).
excludeLastRequest Whether to exclude the records of the last request. The default value is true. Type: boolean (or Expression with resultType boolean).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
sapDataColumnDelimiter The single character that will be used as delimiter passed to SAP RFC as well as splitting the output data retrieved. Type: string (or Expression with resultType string).

SapTableSource

Name Description Value
type Copy source type. "SapTableSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
batchSize Specifies the maximum number of rows that will be retrieved at a time when retrieving data from SAP Table. Type: integer (or Expression with resultType integer).
customRfcReadTableFunctionModule Specifies the custom RFC function module that will be used to read data from SAP Table. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for SAP table read in parallel. Possible values include: "None", "PartitionOnInt", "PartitionOnCalendarYear", "PartitionOnCalendarMonth", "PartitionOnCalendarDate", "PartitionOnTime".
partitionSettings The settings that will be leveraged for SAP table source partitioning. SapTablePartitionSettings
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
rfcTableFields The fields of the SAP table that will be retrieved. For example, column0, column1. Type: string (or Expression with resultType string).
rfcTableOptions The options for the filtering of the SAP Table. For example, COLUMN0 EQ SOME VALUE. Type: string (or Expression with resultType string).
rowCount The number of rows to be retrieved. Type: integer(or Expression with resultType integer).
rowSkips The number of rows that will be skipped. Type: integer (or Expression with resultType integer).
sapDataColumnDelimiter The single character that will be used as delimiter passed to SAP RFC as well as splitting the output data retrieved. Type: string (or Expression with resultType string).

SapTablePartitionSettings

Name Description Value
maxPartitionsNumber The maximum value of partitions the table will be split into. Type: integer (or Expression with resultType string).
partitionColumnName The name of the column that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).
partitionLowerBound The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).
partitionUpperBound The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).

ServiceNowSource

Name Description Value
type Copy source type. "ServiceNowSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

ServiceNowV2Source

Name Description Value
type Copy source type. "ServiceNowV2Source" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
expression Expression to filter data from source. ExpressionV2
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

ExpressionV2

Name Description Value
operands List of nested expressions. ExpressionV2[]
operator Expression operator value Type: string. string
type Type of expressions supported by the system. Type: string. "Binary"
"Constant"
"Field"
"Unary"
value Value for Constant/Field Type: string. string

SharePointOnlineListSource

Name Description Value
type Copy source type. "SharePointOnlineListSource" (required)
httpRequestTimeout The wait time to get a response from SharePoint Online. Default value is 5 minutes (00:05:00). Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
query The OData query to filter the data in SharePoint Online list. For example, "$top=1". Type: string (or Expression with resultType string).

ShopifySource

Name Description Value
type Copy source type. "ShopifySource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

SnowflakeSource

Name Description Value
type Copy source type. "SnowflakeSource" (required)
exportSettings Snowflake export settings. SnowflakeExportCopyCommand (required)
query Snowflake Sql query. Type: string (or Expression with resultType string).

SnowflakeExportCopyCommand

Name Description Value
additionalCopyOptions Additional copy options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": "'HH24:MI:SS.FF'" } SnowflakeExportCopyCommandAdditionalCopyOptions
additionalFormatOptions Additional format options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalFormatOptions": { "OVERWRITE": "TRUE", "MAX_FILE_SIZE": "'FALSE'" } SnowflakeExportCopyCommandAdditionalFormatOptions
type The export setting type. string (required)
{customized property}

SnowflakeExportCopyCommandAdditionalCopyOptions

Name Description Value
{customized property}

SnowflakeExportCopyCommandAdditionalFormatOptions

Name Description Value
{customized property}

SnowflakeV2Source

Name Description Value
type Copy source type. "SnowflakeV2Source" (required)
exportSettings Snowflake export settings. SnowflakeExportCopyCommand (required)
query Snowflake Sql query. Type: string (or Expression with resultType string).

SparkSource

Name Description Value
type Copy source type. "SparkSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

SqlDWSource

Name Description Value
type Copy source type. "SqlDWSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
isolationLevel Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string).
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
sqlReaderQuery SQL Data Warehouse reader query. Type: string (or Expression with resultType string).
sqlReaderStoredProcedureName Name of the stored procedure for a SQL Data Warehouse source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string).
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". Type: object (or Expression with resultType object), itemType: StoredProcedureParameter.

SqlMISource

Name Description Value
type Copy source type. "SqlMISource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
isolationLevel Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string).
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
produceAdditionalTypes Which additional types to produce.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
sqlReaderQuery SQL reader query. Type: string (or Expression with resultType string).
sqlReaderStoredProcedureName Name of the stored procedure for a Azure SQL Managed Instance source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string).
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}".

SqlServerSource

Name Description Value
type Copy source type. "SqlServerSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
isolationLevel Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string).
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
produceAdditionalTypes Which additional types to produce.
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
sqlReaderQuery SQL reader query. Type: string (or Expression with resultType string).
sqlReaderStoredProcedureName Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string).
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}".

SqlSource

Name Description Value
type Copy source type. "SqlSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
isolationLevel Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string).
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
sqlReaderQuery SQL reader query. Type: string (or Expression with resultType string).
sqlReaderStoredProcedureName Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string).
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}".

SquareSource

Name Description Value
type Copy source type. "SquareSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

SybaseSource

Name Description Value
type Copy source type. "SybaseSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query Database query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

TeradataSource

Name Description Value
type Copy source type. "TeradataSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
partitionOption The partition mechanism that will be used for teradata read in parallel. Possible values include: "None", "Hash", "DynamicRange".
partitionSettings The settings that will be leveraged for teradata source partitioning. TeradataPartitionSettings
query Teradata query. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

TeradataPartitionSettings

Name Description Value
partitionColumnName The name of the column that will be used for proceeding range or hash partitioning. Type: string (or Expression with resultType string).
partitionLowerBound The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).
partitionUpperBound The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string).

VerticaSource

Name Description Value
type Copy source type. "VerticaSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

WarehouseSource

Name Description Value
type Copy source type. "WarehouseSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
isolationLevel Specifies the transaction locking behavior for the Microsoft Fabric Warehouse source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string).
partitionOption The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange".
partitionSettings The settings that will be leveraged for Sql source partitioning. SqlPartitionSettings
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
sqlReaderQuery Microsoft Fabric Warehouse reader query. Type: string (or Expression with resultType string).
sqlReaderStoredProcedureName Name of the stored procedure for a Microsoft Fabric Warehouse source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string).
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". Type: object (or Expression with resultType object), itemType: StoredProcedureParameter.

WebSource

Name Description Value
type Copy source type. "WebSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).

XeroSource

Name Description Value
type Copy source type. "XeroSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

XmlSource

Name Description Value
type Copy source type. "XmlSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
formatSettings Xml format settings. XmlReadSettings
storeSettings Xml store settings. StoreReadSettings

XmlReadSettings

Name Description Value
compressionProperties Compression settings. CompressionReadSettings
detectDataType Indicates whether type detection is enabled when reading the xml files. Type: boolean (or Expression with resultType boolean).
namespacePrefixes Namespace uri to prefix mappings to override the prefixes in column names when namespace is enabled, if no prefix is defined for a namespace uri, the prefix of xml element/attribute name in the xml data file will be used. Example: "{"http://www.example.com/xml":"prefix"}" Type: object (or Expression with resultType object).
namespaces Indicates whether namespace is enabled when reading the xml files. Type: boolean (or Expression with resultType boolean).
type The read setting type. string (required)
validationMode Indicates what validation method is used when reading the xml files. Allowed values: 'none', 'xsd', or 'dtd'. Type: string (or Expression with resultType string).
{customized property}

ZohoSource

Name Description Value
type Copy source type. "ZohoSource" (required)
additionalColumns Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects).
query A query to retrieve data from source. Type: string (or Expression with resultType string).
queryTimeout Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

StagingSettings

Name Description Value
enableCompression Specifies whether to use compression when copying data via an interim staging. Default value is false. Type: boolean (or Expression with resultType boolean).
linkedServiceName Staging linked service reference. LinkedServiceReference (required)
path The path to storage for storing the interim data. Type: string (or Expression with resultType string).
{customized property}

CustomActivity

Name Description Value
type Type of activity. "Custom" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Custom activity properties. CustomActivityTypeProperties (required)

CustomActivityTypeProperties

Name Description Value
autoUserSpecification Elevation level and scope for the user, default is nonadmin task. Type: string (or Expression with resultType double).
command Command for custom activity Type: string (or Expression with resultType string).
extendedProperties User defined property bag. There is no restriction on the keys or values that can be used. The user specified custom activity has the full responsibility to consume and interpret the content defined. CustomActivityTypePropertiesExtendedProperties
folderPath Folder path for resource files Type: string (or Expression with resultType string).
referenceObjects Reference objects CustomActivityReferenceObject
resourceLinkedService Resource linked service reference. LinkedServiceReference
retentionTimeInDays The retention time for the files submitted for custom activity. Type: double (or Expression with resultType double).

CustomActivityTypePropertiesExtendedProperties

Name Description Value
{customized property}

CustomActivityReferenceObject

Name Description Value
datasets Dataset references. DatasetReference[]
linkedServices Linked service references. LinkedServiceReference[]

DatabricksNotebookActivity

Name Description Value
type Type of activity. "DatabricksNotebook" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Databricks Notebook activity properties. DatabricksNotebookActivityTypeProperties (required)

DatabricksNotebookActivityTypeProperties

Name Description Value
baseParameters Base parameters to be used for each run of this job.If the notebook takes a parameter that is not specified, the default value from the notebook will be used. DatabricksNotebookActivityTypePropertiesBaseParamete...
libraries A list of libraries to be installed on the cluster that will execute the job. DatabricksNotebookActivityTypePropertiesLibrariesIte...[]
notebookPath The absolute path of the notebook to be run in the Databricks Workspace. This path must begin with a slash. Type: string (or Expression with resultType string).

DatabricksNotebookActivityTypePropertiesBaseParamete...

Name Description Value
{customized property}

DatabricksNotebookActivityTypePropertiesLibrariesIte...

Name Description Value
{customized property}

DatabricksSparkJarActivity

Name Description Value
type Type of activity. "DatabricksSparkJar" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Databricks SparkJar activity properties. DatabricksSparkJarActivityTypeProperties (required)

DatabricksSparkJarActivityTypeProperties

Name Description Value
libraries A list of libraries to be installed on the cluster that will execute the job. DatabricksSparkJarActivityTypePropertiesLibrariesIte...[]
mainClassName The full name of the class containing the main method to be executed. This class must be contained in a JAR provided as a library. Type: string (or Expression with resultType string).
parameters Parameters that will be passed to the main method. any[]

DatabricksSparkJarActivityTypePropertiesLibrariesIte...

Name Description Value
{customized property}

DatabricksSparkPythonActivity

Name Description Value
type Type of activity. "DatabricksSparkPython" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Databricks SparkPython activity properties. DatabricksSparkPythonActivityTypeProperties (required)

DatabricksSparkPythonActivityTypeProperties

Name Description Value
libraries A list of libraries to be installed on the cluster that will execute the job. DatabricksSparkPythonActivityTypePropertiesLibraries...[]
parameters Command line parameters that will be passed to the Python file. any[]
pythonFile The URI of the Python file to be executed. DBFS paths are supported. Type: string (or Expression with resultType string).

DatabricksSparkPythonActivityTypePropertiesLibraries...

Name Description Value
{customized property}

DataLakeAnalyticsUsqlActivity

Name Description Value
type Type of activity. "DataLakeAnalyticsU-SQL" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Data Lake Analytics U-SQL activity properties. DataLakeAnalyticsUsqlActivityTypeProperties (required)

DataLakeAnalyticsUsqlActivityTypeProperties

Name Description Value
compilationMode Compilation mode of U-SQL. Must be one of these values : Semantic, Full and SingleBox. Type: string (or Expression with resultType string).
degreeOfParallelism The maximum number of nodes simultaneously used to run the job. Default value is 1. Type: integer (or Expression with resultType integer), minimum: 1.
parameters Parameters for U-SQL job request. DataLakeAnalyticsUsqlActivityTypePropertiesParameter...
priority Determines which jobs out of all that are queued should be selected to run first. The lower the number, the higher the priority. Default value is 1000. Type: integer (or Expression with resultType integer), minimum: 1.
runtimeVersion Runtime version of the U-SQL engine to use. Type: string (or Expression with resultType string).
scriptLinkedService Script linked service reference. LinkedServiceReference (required)
scriptPath Case-sensitive path to folder that contains the U-SQL script. Type: string (or Expression with resultType string).

DataLakeAnalyticsUsqlActivityTypePropertiesParameter...

Name Description Value
{customized property}

DeleteActivity

Name Description Value
type Type of activity. "Delete" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Delete activity properties. DeleteActivityTypeProperties (required)

DeleteActivityTypeProperties

Name Description Value
dataset Delete activity dataset reference. DatasetReference (required)
enableLogging Whether to record detailed logs of delete-activity execution. Default value is false. Type: boolean (or Expression with resultType boolean).
logStorageSettings Log storage settings customer need to provide when enableLogging is true. LogStorageSettings
maxConcurrentConnections The max concurrent connections to connect data source at the same time. int

Constraints:
Min value = 1
recursive If true, files or sub-folders under current folder path will be deleted recursively. Default is false. Type: boolean (or Expression with resultType boolean).
storeSettings Delete activity store settings. StoreReadSettings

ExecuteDataFlowActivity

Name Description Value
type Type of activity. "ExecuteDataFlow" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Execute data flow activity properties. ExecuteDataFlowActivityTypeProperties (required)

ExecuteDataFlowActivityTypeProperties

Name Description Value
compute Compute properties for data flow activity. ExecuteDataFlowActivityTypePropertiesCompute
continueOnError Continue on error setting used for data flow execution. Enables processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean)
dataFlow Data flow reference. DataFlowReference (required)
integrationRuntime The integration runtime reference. IntegrationRuntimeReference
runConcurrently Concurrent run setting used for data flow execution. Allows sinks with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean)
sourceStagingConcurrency Specify number of parallel staging for sources applicable to the sink. Type: integer (or Expression with resultType integer)
staging Staging info for execute data flow activity. DataFlowStagingInfo
traceLevel Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string)

ExecuteDataFlowActivityTypePropertiesCompute

Name Description Value
computeType Compute type of the cluster which will execute data flow job. Possible values include: 'General', 'MemoryOptimized', 'ComputeOptimized'. Type: string (or Expression with resultType string)
coreCount Core count of the cluster which will execute data flow job. Supported values are: 8, 16, 32, 48, 80, 144 and 272. Type: integer (or Expression with resultType integer)

DataFlowReference

Name Description Value
datasetParameters Reference data flow parameters from dataset.
parameters Data flow parameters ParameterValueSpecification
referenceName Reference data flow name. string (required)
type Data flow reference type. "DataFlowReference" (required)
{customized property}

IntegrationRuntimeReference

Name Description Value
parameters Arguments for integration runtime. ParameterValueSpecification
referenceName Reference integration runtime name. string (required)
type Type of integration runtime. "IntegrationRuntimeReference" (required)

DataFlowStagingInfo

Name Description Value
folderPath Folder path for staging blob. Type: string (or Expression with resultType string)
linkedService Staging linked service reference. LinkedServiceReference

ExecutePipelineActivity

Name Description Value
type Type of activity. "ExecutePipeline" (required)
policy Execute pipeline activity policy. ExecutePipelineActivityPolicy
typeProperties Execute pipeline activity properties. ExecutePipelineActivityTypeProperties (required)

ExecutePipelineActivityPolicy

Name Description Value
secureInput When set to true, Input from activity is considered as secure and will not be logged to monitoring. bool
{customized property}

ExecutePipelineActivityTypeProperties

Name Description Value
parameters Pipeline parameters. ParameterValueSpecification
pipeline Pipeline reference. PipelineReference (required)
waitOnCompletion Defines whether activity execution will wait for the dependent pipeline execution to finish. Default is false. bool

PipelineReference

Name Description Value
name Reference name. string
referenceName Reference pipeline name. string (required)
type Pipeline reference type. "PipelineReference" (required)

ExecuteSsisPackageActivity

Name Description Value
type Type of activity. "ExecuteSSISPackage" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Execute SSIS package activity properties. ExecuteSsisPackageActivityTypeProperties (required)

ExecuteSsisPackageActivityTypeProperties

Name Description Value
connectVia The integration runtime reference. IntegrationRuntimeReference (required)
environmentPath The environment path to execute the SSIS package. Type: string (or Expression with resultType string).
executionCredential The package execution credential. SsisExecutionCredential
loggingLevel The logging level of SSIS package execution. Type: string (or Expression with resultType string).
logLocation SSIS package execution log location. SsisLogLocation
packageConnectionManagers The package level connection managers to execute the SSIS package. ExecuteSsisPackageActivityTypePropertiesPackageConne...
packageLocation SSIS package location. SsisPackageLocation (required)
packageParameters The package level parameters to execute the SSIS package. ExecuteSsisPackageActivityTypePropertiesPackageParam...
projectConnectionManagers The project level connection managers to execute the SSIS package. ExecuteSsisPackageActivityTypePropertiesProjectConne...
projectParameters The project level parameters to execute the SSIS package. ExecuteSsisPackageActivityTypePropertiesProjectParam...
propertyOverrides The property overrides to execute the SSIS package. ExecuteSsisPackageActivityTypePropertiesPropertyOver...
runtime Specifies the runtime to execute SSIS package. The value should be "x86" or "x64". Type: string (or Expression with resultType string).

SsisExecutionCredential

Name Description Value
domain Domain for windows authentication. Type: string (or Expression with resultType string).
password Password for windows authentication. SecureString (required)
userName UseName for windows authentication. Type: string (or Expression with resultType string).

SecureString

Name Description Value
type Type of the secret. string (required)
value Value of secure string. string (required)

SsisLogLocation

Name Description Value
logPath The SSIS package execution log path. Type: string (or Expression with resultType string).
type The type of SSIS log location. "File" (required)
typeProperties SSIS package execution log location properties. SsisLogLocationTypeProperties (required)

SsisLogLocationTypeProperties

Name Description Value
accessCredential The package execution log access credential. SsisAccessCredential
logRefreshInterval Specifies the interval to refresh log. The default interval is 5 minutes. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

SsisAccessCredential

Name Description Value
domain Domain for windows authentication. Type: string (or Expression with resultType string).
password Password for windows authentication. SecretBase (required)
userName UseName for windows authentication. Type: string (or Expression with resultType string).

SecretBase

Name Description Value
type Set the object type AzureKeyVaultSecret
SecureString (required)

AzureKeyVaultSecretReference

Name Description Value
secretName The name of the secret in Azure Key Vault. Type: string (or Expression with resultType string).
secretVersion The version of the secret in Azure Key Vault. The default value is the latest version of the secret. Type: string (or Expression with resultType string).
store The Azure Key Vault linked service reference. LinkedServiceReference (required)
type Type of the secret. string (required)

ExecuteSsisPackageActivityTypePropertiesPackageConne...

Name Description Value
{customized property} SsisConnectionManager

SsisConnectionManager

Name Description Value
{customized property} SsisExecutionParameter

SsisExecutionParameter

Name Description Value
value SSIS package execution parameter value. Type: string (or Expression with resultType string).

SsisPackageLocation

Name Description Value
packagePath The SSIS package path. Type: string (or Expression with resultType string).
type The type of SSIS package location. "File"
"InlinePackage"
"PackageStore"
"SSISDB"
typeProperties SSIS package location properties. SsisPackageLocationTypeProperties

SsisPackageLocationTypeProperties

Name Description Value
accessCredential The package access credential. SsisAccessCredential
childPackages The embedded child package list. SsisChildPackage[]
configurationAccessCredential The configuration file access credential. SsisAccessCredential
configurationPath The configuration file of the package execution. Type: string (or Expression with resultType string).
packageContent The embedded package content. Type: string (or Expression with resultType string).
packageLastModifiedDate The embedded package last modified date. string
packageName The package name. string
packagePassword Password of the package. SecretBase

SsisChildPackage

Name Description Value
packageContent Content for embedded child package. Type: string (or Expression with resultType string).
packageLastModifiedDate Last modified date for embedded child package. string
packageName Name for embedded child package. string
packagePath Path for embedded child package. Type: string (or Expression with resultType string).

ExecuteSsisPackageActivityTypePropertiesPackageParam...

Name Description Value
{customized property} SsisExecutionParameter

ExecuteSsisPackageActivityTypePropertiesProjectConne...

Name Description Value
{customized property} SsisConnectionManager

ExecuteSsisPackageActivityTypePropertiesProjectParam...

Name Description Value
{customized property} SsisExecutionParameter

ExecuteSsisPackageActivityTypePropertiesPropertyOver...

Name Description Value
{customized property} SsisPropertyOverride

SsisPropertyOverride

Name Description Value
isSensitive Whether SSIS package property override value is sensitive data. Value will be encrypted in SSISDB if it is true bool
value SSIS package property override value. Type: string (or Expression with resultType string).

ExecuteWranglingDataflowActivity

Name Description Value
type Type of activity. "ExecuteWranglingDataflow" (required)
policy Activity policy. ActivityPolicy
typeProperties Execute power query activity properties. ExecutePowerQueryActivityTypeProperties (required)

ExecutePowerQueryActivityTypeProperties

Name Description Value
compute Compute properties for data flow activity. ExecuteDataFlowActivityTypePropertiesCompute
continueOnError Continue on error setting used for data flow execution. Enables processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean)
dataFlow Data flow reference. DataFlowReference (required)
integrationRuntime The integration runtime reference. IntegrationRuntimeReference
queries List of mapping for Power Query mashup query to sink dataset(s). PowerQuerySinkMapping[]
runConcurrently Concurrent run setting used for data flow execution. Allows sinks with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean)
sinks (Deprecated. Please use Queries). List of Power Query activity sinks mapped to a queryName. ExecutePowerQueryActivityTypePropertiesSinks
sourceStagingConcurrency Specify number of parallel staging for sources applicable to the sink. Type: integer (or Expression with resultType integer)
staging Staging info for execute data flow activity. DataFlowStagingInfo
traceLevel Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string)

PowerQuerySinkMapping

Name Description Value
dataflowSinks List of sinks mapped to Power Query mashup query. PowerQuerySink[]
queryName Name of the query in Power Query mashup document. string

PowerQuerySink

Name Description Value
dataset Dataset reference. DatasetReference
description Transformation description. string
flowlet Flowlet Reference DataFlowReference
linkedService Linked service reference. LinkedServiceReference
name Transformation name. string (required)
rejectedDataLinkedService Rejected data linked service reference. LinkedServiceReference
schemaLinkedService Schema linked service reference. LinkedServiceReference
script sink script. string

ExecutePowerQueryActivityTypePropertiesSinks

Name Description Value
{customized property} PowerQuerySink

FailActivity

Name Description Value
type Type of activity. "Fail" (required)
typeProperties Fail activity properties. FailActivityTypeProperties (required)

FailActivityTypeProperties

Name Description Value
errorCode The error code that categorizes the error type of the Fail activity. It can be dynamic content that's evaluated to a non empty/blank string at runtime. Type: string (or Expression with resultType string).
message The error message that surfaced in the Fail activity. It can be dynamic content that's evaluated to a non empty/blank string at runtime. Type: string (or Expression with resultType string).

FilterActivity

Name Description Value
type Type of activity. "Filter" (required)
typeProperties Filter activity properties. FilterActivityTypeProperties (required)

FilterActivityTypeProperties

Name Description Value
condition Condition to be used for filtering the input. Expression (required)
items Input array on which filter should be applied. Expression (required)

Expression

Name Description Value
type Expression type. "Expression" (required)
value Expression value. string (required)

ForEachActivity

Name Description Value
type Type of activity. "ForEach" (required)
typeProperties ForEach activity properties. ForEachActivityTypeProperties (required)

ForEachActivityTypeProperties

Name Description Value
activities List of activities to execute . Activity[] (required)
batchCount Batch count to be used for controlling the number of parallel execution (when isSequential is set to false). int

Constraints:
Max value = 50
isSequential Should the loop be executed in sequence or in parallel (max 50) bool
items Collection to iterate. Expression (required)

GetMetadataActivity

Name Description Value
type Type of activity. "GetMetadata" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties GetMetadata activity properties. GetMetadataActivityTypeProperties (required)

GetMetadataActivityTypeProperties

Name Description Value
dataset GetMetadata activity dataset reference. DatasetReference (required)
fieldList Fields of metadata to get from dataset. any[]
formatSettings GetMetadata activity format settings. FormatReadSettings
storeSettings GetMetadata activity store settings. StoreReadSettings

FormatReadSettings

Name Description Value
type Set the object type BinaryReadSettings
DelimitedTextReadSettings
JsonReadSettings
ParquetReadSettings
XmlReadSettings (required)

HDInsightHiveActivity

Name Description Value
type Type of activity. "HDInsightHive" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties HDInsight Hive activity properties. HDInsightHiveActivityTypeProperties (required)

HDInsightHiveActivityTypeProperties

Name Description Value
arguments User specified arguments to HDInsightActivity. any[]
defines Allows user to specify defines for Hive job request. HDInsightHiveActivityTypePropertiesDefines
getDebugInfo Debug info option. "Always"
"Failure"
"None"
queryTimeout Query timeout value (in minutes). Effective when the HDInsight cluster is with ESP (Enterprise Security Package) int
scriptLinkedService Script linked service reference. LinkedServiceReference
scriptPath Script path. Type: string (or Expression with resultType string).
storageLinkedServices Storage linked service references. LinkedServiceReference[]
variables User specified arguments under hivevar namespace. HDInsightHiveActivityTypePropertiesVariables

HDInsightHiveActivityTypePropertiesDefines

Name Description Value
{customized property}

HDInsightHiveActivityTypePropertiesVariables

Name Description Value
{customized property}

HDInsightMapReduceActivity

Name Description Value
type Type of activity. "HDInsightMapReduce" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties HDInsight MapReduce activity properties. HDInsightMapReduceActivityTypeProperties (required)

HDInsightMapReduceActivityTypeProperties

Name Description Value
arguments User specified arguments to HDInsightActivity. any[]
className Class name. Type: string (or Expression with resultType string).
defines Allows user to specify defines for the MapReduce job request. HDInsightMapReduceActivityTypePropertiesDefines
getDebugInfo Debug info option. "Always"
"Failure"
"None"
jarFilePath Jar path. Type: string (or Expression with resultType string).
jarLibs Jar libs. any[]
jarLinkedService Jar linked service reference. LinkedServiceReference
storageLinkedServices Storage linked service references. LinkedServiceReference[]

HDInsightMapReduceActivityTypePropertiesDefines

Name Description Value
{customized property}

HDInsightPigActivity

Name Description Value
type Type of activity. "HDInsightPig" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties HDInsight Pig activity properties. HDInsightPigActivityTypeProperties (required)

HDInsightPigActivityTypeProperties

Name Description Value
arguments User specified arguments to HDInsightActivity. Type: array (or Expression with resultType array).
defines Allows user to specify defines for Pig job request. HDInsightPigActivityTypePropertiesDefines
getDebugInfo Debug info option. "Always"
"Failure"
"None"
scriptLinkedService Script linked service reference. LinkedServiceReference
scriptPath Script path. Type: string (or Expression with resultType string).
storageLinkedServices Storage linked service references. LinkedServiceReference[]

HDInsightPigActivityTypePropertiesDefines

Name Description Value
{customized property}

HDInsightSparkActivity

Name Description Value
type Type of activity. "HDInsightSpark" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties HDInsight spark activity properties. HDInsightSparkActivityTypeProperties (required)

HDInsightSparkActivityTypeProperties

Name Description Value
arguments The user-specified arguments to HDInsightSparkActivity. any[]
className The application's Java/Spark main class. string
entryFilePath The relative path to the root folder of the code/package to be executed. Type: string (or Expression with resultType string).
getDebugInfo Debug info option. "Always"
"Failure"
"None"
proxyUser The user to impersonate that will execute the job. Type: string (or Expression with resultType string).
rootPath The root path in 'sparkJobLinkedService' for all the job’s files. Type: string (or Expression with resultType string).
sparkConfig Spark configuration property. HDInsightSparkActivityTypePropertiesSparkConfig
sparkJobLinkedService The storage linked service for uploading the entry file and dependencies, and for receiving logs. LinkedServiceReference

HDInsightSparkActivityTypePropertiesSparkConfig

Name Description Value
{customized property}

HDInsightStreamingActivity

Name Description Value
type Type of activity. "HDInsightStreaming" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties HDInsight streaming activity properties. HDInsightStreamingActivityTypeProperties (required)

HDInsightStreamingActivityTypeProperties

Name Description Value
arguments User specified arguments to HDInsightActivity. any[]
combiner Combiner executable name. Type: string (or Expression with resultType string).
commandEnvironment Command line environment values. any[]
defines Allows user to specify defines for streaming job request. HDInsightStreamingActivityTypePropertiesDefines
fileLinkedService Linked service reference where the files are located. LinkedServiceReference
filePaths Paths to streaming job files. Can be directories. any[] (required)
getDebugInfo Debug info option. "Always"
"Failure"
"None"
input Input blob path. Type: string (or Expression with resultType string).
mapper Mapper executable name. Type: string (or Expression with resultType string).
output Output blob path. Type: string (or Expression with resultType string).
reducer Reducer executable name. Type: string (or Expression with resultType string).
storageLinkedServices Storage linked service references. LinkedServiceReference[]

HDInsightStreamingActivityTypePropertiesDefines

Name Description Value
{customized property}

IfConditionActivity

Name Description Value
type Type of activity. "IfCondition" (required)
typeProperties IfCondition activity properties. IfConditionActivityTypeProperties (required)

IfConditionActivityTypeProperties

Name Description Value
expression An expression that would evaluate to Boolean. This is used to determine the block of activities (ifTrueActivities or ifFalseActivities) that will be executed. Expression (required)
ifFalseActivities List of activities to execute if expression is evaluated to false. This is an optional property and if not provided, the activity will exit without any action. Activity[]
ifTrueActivities List of activities to execute if expression is evaluated to true. This is an optional property and if not provided, the activity will exit without any action. Activity[]

LookupActivity

Name Description Value
type Type of activity. "Lookup" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Lookup activity properties. LookupActivityTypeProperties (required)

LookupActivityTypeProperties

Name Description Value
dataset Lookup activity dataset reference. DatasetReference (required)
firstRowOnly Whether to return first row or all rows. Default value is true. Type: boolean (or Expression with resultType boolean).
source Dataset-specific source properties, same as copy activity source. CopySource (required)

ScriptActivity

Name Description Value
type Type of activity. "Script" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Script activity properties. ScriptActivityTypeProperties (required)

ScriptActivityTypeProperties

Name Description Value
logSettings Log settings of script activity. ScriptActivityTypePropertiesLogSettings
scriptBlockExecutionTimeout ScriptBlock execution timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).
scripts Array of script blocks. Type: array. ScriptActivityScriptBlock[]

ScriptActivityTypePropertiesLogSettings

Name Description Value
logDestination The destination of logs. Type: string. "ActivityOutput"
"ExternalStore" (required)
logLocationSettings Log location settings customer needs to provide when enabling log. LogLocationSettings

ScriptActivityScriptBlock

Name Description Value
parameters Array of script parameters. Type: array. ScriptActivityParameter[]
text The query text. Type: string (or Expression with resultType string).
type The type of the query. Type: string. "NonQuery"
"Query" (required)

ScriptActivityParameter

Name Description Value
direction The direction of the parameter. "Input"
"InputOutput"
"Output"
name The name of the parameter. Type: string (or Expression with resultType string).
size The size of the output direction parameter. int
type The type of the parameter. "Boolean"
"DateTime"
"DateTimeOffset"
"Decimal"
"Double"
"Guid"
"Int16"
"Int32"
"Int64"
"Single"
"String"
"Timespan"
value The value of the parameter. Type: string (or Expression with resultType string).

SetVariableActivity

Name Description Value
type Type of activity. "SetVariable" (required)
policy Activity policy. SecureInputOutputPolicy
typeProperties Set Variable activity properties. SetVariableActivityTypeProperties (required)

SecureInputOutputPolicy

Name Description Value
secureInput When set to true, Input from activity is considered as secure and will not be logged to monitoring. bool
secureOutput When set to true, Output from activity is considered as secure and will not be logged to monitoring. bool

SetVariableActivityTypeProperties

Name Description Value
setSystemVariable If set to true, it sets the pipeline run return value. bool
value Value to be set. Could be a static value or Expression.
variableName Name of the variable whose value needs to be set. string

SynapseSparkJobDefinitionActivity

Name Description Value
type Type of activity. "SparkJob" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Execute spark job activity properties. SynapseSparkJobActivityTypeProperties (required)

SynapseSparkJobActivityTypeProperties

Name Description Value
args User specified arguments to SynapseSparkJobDefinitionActivity. any[]
className The fully-qualified identifier or the main class that is in the main definition file, which will override the 'className' of the spark job definition you provide. Type: string (or Expression with resultType string).
conf Spark configuration properties, which will override the 'conf' of the spark job definition you provide.
configurationType The type of the spark config. "Artifact"
"Customized"
"Default"
driverSize Number of core and memory to be used for driver allocated in the specified Spark pool for the job, which will be used for overriding 'driverCores' and 'driverMemory' of the spark job definition you provide. Type: string (or Expression with resultType string).
executorSize Number of core and memory to be used for executors allocated in the specified Spark pool for the job, which will be used for overriding 'executorCores' and 'executorMemory' of the spark job definition you provide. Type: string (or Expression with resultType string).
file The main file used for the job, which will override the 'file' of the spark job definition you provide. Type: string (or Expression with resultType string).
files (Deprecated. Please use pythonCodeReference and filesV2) Additional files used for reference in the main definition file, which will override the 'files' of the spark job definition you provide. any[]
filesV2 Additional files used for reference in the main definition file, which will override the 'jars' and 'files' of the spark job definition you provide. any[]
numExecutors Number of executors to launch for this job, which will override the 'numExecutors' of the spark job definition you provide. Type: integer (or Expression with resultType integer).
pythonCodeReference Additional python code files used for reference in the main definition file, which will override the 'pyFiles' of the spark job definition you provide. any[]
scanFolder Scanning subfolders from the root folder of the main definition file, these files will be added as reference files. The folders named 'jars', 'pyFiles', 'files' or 'archives' will be scanned, and the folders name are case sensitive. Type: boolean (or Expression with resultType boolean).
sparkConfig Spark configuration property. SynapseSparkJobActivityTypePropertiesSparkConfig
sparkJob Synapse spark job reference. SynapseSparkJobReference (required)
targetBigDataPool The name of the big data pool which will be used to execute the spark batch job, which will override the 'targetBigDataPool' of the spark job definition you provide. BigDataPoolParametrizationReference
targetSparkConfiguration The spark configuration of the spark job. SparkConfigurationParametrizationReference

SynapseSparkJobActivityTypePropertiesSparkConfig

Name Description Value
{customized property}

SynapseSparkJobReference

Name Description Value
referenceName Reference spark job name. Expression with resultType string.
type Synapse spark job reference type. "SparkJobDefinitionReference" (required)

BigDataPoolParametrizationReference

Name Description Value
referenceName Reference big data pool name. Type: string (or Expression with resultType string).
type Big data pool reference type. "BigDataPoolReference" (required)

SparkConfigurationParametrizationReference

Name Description Value
referenceName Reference spark configuration name. Type: string (or Expression with resultType string).
type Spark configuration reference type. "SparkConfigurationReference" (required)

SqlServerStoredProcedureActivity

Name Description Value
type Type of activity. "SqlServerStoredProcedure" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties SQL stored procedure activity properties. SqlServerStoredProcedureActivityTypeProperties (required)

SqlServerStoredProcedureActivityTypeProperties

Name Description Value
storedProcedureName Stored procedure name. Type: string (or Expression with resultType string).
storedProcedureParameters Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}".

SwitchActivity

Name Description Value
type Type of activity. "Switch" (required)
typeProperties Switch activity properties. SwitchActivityTypeProperties (required)

SwitchActivityTypeProperties

Name Description Value
cases List of cases that correspond to expected values of the 'on' property. This is an optional property and if not provided, the activity will execute activities provided in defaultActivities. SwitchCase[]
defaultActivities List of activities to execute if no case condition is satisfied. This is an optional property and if not provided, the activity will exit without any action. Activity[]
on An expression that would evaluate to a string or integer. This is used to determine the block of activities in cases that will be executed. Expression (required)

SwitchCase

Name Description Value
activities List of activities to execute for satisfied case condition. Activity[]
value Expected value that satisfies the expression result of the 'on' property. string

SynapseNotebookActivity

Name Description Value
type Type of activity. "SynapseNotebook" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Execute Synapse notebook activity properties. SynapseNotebookActivityTypeProperties (required)

SynapseNotebookActivityTypeProperties

Name Description Value
conf Spark configuration properties, which will override the 'conf' of the notebook you provide.
configurationType The type of the spark config. "Artifact"
"Customized"
"Default"
driverSize Number of core and memory to be used for driver allocated in the specified Spark pool for the session, which will be used for overriding 'driverCores' and 'driverMemory' of the notebook you provide. Type: string (or Expression with resultType string).
executorSize Number of core and memory to be used for executors allocated in the specified Spark pool for the session, which will be used for overriding 'executorCores' and 'executorMemory' of the notebook you provide. Type: string (or Expression with resultType string).
notebook Synapse notebook reference. SynapseNotebookReference (required)
numExecutors Number of executors to launch for this session, which will override the 'numExecutors' of the notebook you provide. Type: integer (or Expression with resultType integer).
parameters Notebook parameters. SynapseNotebookActivityTypePropertiesParameters
sparkConfig Spark configuration property. SynapseNotebookActivityTypePropertiesSparkConfig
sparkPool The name of the big data pool which will be used to execute the notebook. BigDataPoolParametrizationReference
targetSparkConfiguration The spark configuration of the spark job. SparkConfigurationParametrizationReference

SynapseNotebookReference

Name Description Value
referenceName Reference notebook name. Type: string (or Expression with resultType string).
type Synapse notebook reference type. "NotebookReference" (required)

SynapseNotebookActivityTypePropertiesParameters

Name Description Value
{customized property} NotebookParameter

NotebookParameter

Name Description Value
type Notebook parameter type. "bool"
"float"
"int"
"string"
value Notebook parameter value. Type: string (or Expression with resultType string).

SynapseNotebookActivityTypePropertiesSparkConfig

Name Description Value
{customized property}

UntilActivity

Name Description Value
type Type of activity. "Until" (required)
typeProperties Until activity properties. UntilActivityTypeProperties (required)

UntilActivityTypeProperties

Name Description Value
activities List of activities to execute. Activity[] (required)
expression An expression that would evaluate to Boolean. The loop will continue until this expression evaluates to true Expression (required)
timeout Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

ValidationActivity

Name Description Value
type Type of activity. "Validation" (required)
typeProperties Validation activity properties. ValidationActivityTypeProperties (required)

ValidationActivityTypeProperties

Name Description Value
childItems Can be used if dataset points to a folder. If set to true, the folder must have at least one file. If set to false, the folder must be empty. Type: boolean (or Expression with resultType boolean).
dataset Validation activity dataset reference. DatasetReference (required)
minimumSize Can be used if dataset points to a file. The file must be greater than or equal in size to the value specified. Type: integer (or Expression with resultType integer).
sleep A delay in seconds between validation attempts. If no value is specified, 10 seconds will be used as the default. Type: integer (or Expression with resultType integer).
timeout Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).

WaitActivity

Name Description Value
type Type of activity. "Wait" (required)
typeProperties Wait activity properties. WaitActivityTypeProperties (required)

WaitActivityTypeProperties

Name Description Value
waitTimeInSeconds Duration in seconds. Type: integer (or Expression with resultType integer).

WebActivity

Name Description Value
type Type of activity. "WebActivity" (required)
linkedServiceName Linked service reference. LinkedServiceReference
policy Activity policy. ActivityPolicy
typeProperties Web activity properties. WebActivityTypeProperties (required)

WebActivityTypeProperties

Name Description Value
authentication Authentication method used for calling the endpoint. WebActivityAuthentication
body Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string).
connectVia The integration runtime reference. IntegrationRuntimeReference
datasets List of datasets passed to web endpoint. DatasetReference[]
disableCertValidation When set to true, Certificate validation will be disabled. bool
headers Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). WebActivityTypePropertiesHeaders
httpRequestTimeout Timeout for the HTTP request to get a response. Format is in TimeSpan (hh:mm:ss). This value is the timeout to get a response, not the activity timeout. The default value is 00:01:00 (1 minute). The range is from 1 to 10 minutes
linkedServices List of linked services passed to web endpoint. LinkedServiceReference[]
method Rest API method for target endpoint. "DELETE"
"GET"
"POST"
"PUT" (required)
turnOffAsync Option to disable invoking HTTP GET on location given in response header of a HTTP 202 Response. If set true, it stops invoking HTTP GET on http location given in response header. If set false then continues to invoke HTTP GET call on location given in http response headers. bool
url Web activity target endpoint and path. Type: string (or Expression with resultType string).

WebActivityAuthentication

Name Description Value
credential The credential reference containing authentication information. CredentialReference
password Password for the PFX file or basic authentication / Secret when used for ServicePrincipal SecretBase
pfx Base64-encoded contents of a PFX file or Certificate when used for ServicePrincipal SecretBase
resource Resource for which Azure Auth token will be requested when using MSI Authentication. Type: string (or Expression with resultType string).
type Web activity authentication (Basic/ClientCertificate/MSI/ServicePrincipal) string
username Web activity authentication user name for basic authentication or ClientID when used for ServicePrincipal. Type: string (or Expression with resultType string).
userTenant TenantId for which Azure Auth token will be requested when using ServicePrincipal Authentication. Type: string (or Expression with resultType string).

CredentialReference

Name Description Value
referenceName Reference credential name. string (required)
type Credential reference type. "CredentialReference" (required)
{customized property}

WebActivityTypePropertiesHeaders

Name Description Value
{customized property} string

WebHookActivity

Name Description Value
type Type of activity. "WebHook" (required)
policy Activity policy. SecureInputOutputPolicy
typeProperties WebHook activity properties. WebHookActivityTypeProperties (required)

WebHookActivityTypeProperties

Name Description Value
authentication Authentication method used for calling the endpoint. WebActivityAuthentication
body Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string).
headers Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). WebHookActivityTypePropertiesHeaders
method Rest API method for target endpoint. "POST" (required)
reportStatusOnCallBack When set to true, statusCode, output and error in callback request body will be consumed by activity. The activity can be marked as failed by setting statusCode >= 400 in callback request. Default is false. Type: boolean (or Expression with resultType boolean).
timeout The timeout within which the webhook should be called back. If there is no value specified, it defaults to 10 minutes. Type: string. Pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). string
url WebHook activity target endpoint and path. Type: string (or Expression with resultType string).

WebHookActivityTypePropertiesHeaders

Name Description Value
{customized property} string

PipelineFolder

Name Description Value
name The name of the folder that this Pipeline is in. string

ParameterDefinitionSpecification

Name Description Value
{customized property} ParameterSpecification

ParameterSpecification

Name Description Value
defaultValue Default value of parameter.
type Parameter type. "Array"
"Bool"
"Float"
"Int"
"Object"
"SecureString"
"String" (required)

PipelinePolicy

Name Description Value
elapsedTimeMetric Pipeline ElapsedTime Metric Policy. PipelineElapsedTimeMetricPolicy

PipelineElapsedTimeMetricPolicy

Name Description Value
duration TimeSpan value, after which an Azure Monitoring Metric is fired.

PipelineRunDimensions

Name Description Value
{customized property}

VariableDefinitionSpecification

Name Description Value
{customized property} VariableSpecification

VariableSpecification

Name Description Value
defaultValue Default value of variable.
type Variable type. "Array"
"Bool"
"String" (required)