Microsoft.DataFactory factories/pipelines 2018-06-01
Bicep resource definition
The factories/pipelines resource type can be deployed with operations that target:
- Resource groups - See resource group deployment commands
For a list of changed properties in each API version, see change log.
Resource format
To create a Microsoft.DataFactory/factories/pipelines resource, add the following Bicep to your template.
resource symbolicname 'Microsoft.DataFactory/factories/pipelines@2018-06-01' = {
name: 'string'
parent: resourceSymbolicName
properties: {
activities: [
{
dependsOn: [
{
activity: 'string'
dependencyConditions: [
'string'
]
{customized property}: any()
}
]
description: 'string'
name: 'string'
onInactiveMarkAs: 'string'
state: 'string'
userProperties: [
{
name: 'string'
value: any()
}
]
type: 'string'
// For remaining properties, see Activity objects
}
]
annotations: [
any
]
concurrency: int
description: 'string'
folder: {
name: 'string'
}
parameters: {
{customized property}: {
defaultValue: any()
type: 'string'
}
}
policy: {
elapsedTimeMetric: {
duration: any()
}
}
runDimensions: {
{customized property}: any()
}
variables: {
{customized property}: {
defaultValue: any()
type: 'string'
}
}
}
}
Activity objects
Set the type property to specify the type of object.
For AppendVariable, use:
type: 'AppendVariable'
typeProperties: {
value: any()
variableName: 'string'
}
For AzureDataExplorerCommand, use:
type: 'AzureDataExplorerCommand'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
command: any()
commandTimeout: any()
}
For AzureFunctionActivity, use:
type: 'AzureFunctionActivity'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
body: any()
functionName: any()
headers: {
{customized property}: 'string'
}
method: 'string'
}
For AzureMLBatchExecution, use:
type: 'AzureMLBatchExecution'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
globalParameters: {
{customized property}: any()
}
webServiceInputs: {
{customized property}: {
filePath: any()
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
}
}
webServiceOutputs: {
{customized property}: {
filePath: any()
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
}
}
}
For AzureMLExecutePipeline, use:
type: 'AzureMLExecutePipeline'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
continueOnStepFailure: any()
dataPathAssignments: any()
experimentName: any()
mlParentRunId: any()
mlPipelineEndpointId: any()
mlPipelineId: any()
mlPipelineParameters: any()
version: any()
}
For AzureMLUpdateResource, use:
type: 'AzureMLUpdateResource'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
trainedModelFilePath: any()
trainedModelLinkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
trainedModelName: any()
}
For Copy, use:
type: 'Copy'
inputs: [
{
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'DatasetReference'
}
]
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
outputs: [
{
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'DatasetReference'
}
]
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
dataIntegrationUnits: any()
enableSkipIncompatibleRow: any()
enableStaging: any()
logSettings: {
copyActivityLogSettings: {
enableReliableLogging: any()
logLevel: any()
}
enableCopyActivityLog: any()
logLocationSettings: {
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
path: any()
}
}
logStorageSettings: {
enableReliableLogging: any()
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
logLevel: any()
path: any()
{customized property}: any()
}
parallelCopies: any()
preserve: [
any
]
preserveRules: [
any
]
redirectIncompatibleRowSettings: {
linkedServiceName: any()
path: any()
{customized property}: any()
}
sink: {
disableMetricsCollection: any()
maxConcurrentConnections: any()
sinkRetryCount: any()
sinkRetryWait: any()
writeBatchSize: any()
writeBatchTimeout: any()
type: 'string'
// For remaining properties, see CopySink objects
}
skipErrorFile: {
dataInconsistency: any()
fileMissing: any()
}
source: {
disableMetricsCollection: any()
maxConcurrentConnections: any()
sourceRetryCount: any()
sourceRetryWait: any()
type: 'string'
// For remaining properties, see CopySource objects
}
stagingSettings: {
enableCompression: any()
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
path: any()
{customized property}: any()
}
translator: any()
validateDataConsistency: any()
}
For Custom, use:
type: 'Custom'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
autoUserSpecification: any()
command: any()
extendedProperties: {
{customized property}: any()
}
folderPath: any()
referenceObjects: {
datasets: [
{
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'DatasetReference'
}
]
linkedServices: [
{
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
]
}
resourceLinkedService: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
retentionTimeInDays: any()
}
For DatabricksNotebook, use:
type: 'DatabricksNotebook'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
baseParameters: {
{customized property}: any()
}
libraries: [
{
{customized property}: any()
}
]
notebookPath: any()
}
For DatabricksSparkJar, use:
type: 'DatabricksSparkJar'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
libraries: [
{
{customized property}: any()
}
]
mainClassName: any()
parameters: [
any
]
}
For DatabricksSparkPython, use:
type: 'DatabricksSparkPython'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
libraries: [
{
{customized property}: any()
}
]
parameters: [
any
]
pythonFile: any()
}
For DataLakeAnalyticsU-SQL, use:
type: 'DataLakeAnalyticsU-SQL'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
compilationMode: any()
degreeOfParallelism: any()
parameters: {
{customized property}: any()
}
priority: any()
runtimeVersion: any()
scriptLinkedService: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
scriptPath: any()
}
For Delete, use:
type: 'Delete'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
dataset: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'DatasetReference'
}
enableLogging: any()
logStorageSettings: {
enableReliableLogging: any()
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
logLevel: any()
path: any()
{customized property}: any()
}
maxConcurrentConnections: int
recursive: any()
storeSettings: {
disableMetricsCollection: any()
maxConcurrentConnections: any()
type: 'string'
// For remaining properties, see StoreReadSettings objects
}
}
For ExecuteDataFlow, use:
type: 'ExecuteDataFlow'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
compute: {
computeType: any()
coreCount: any()
}
continueOnError: any()
dataFlow: {
datasetParameters: any()
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'DataFlowReference'
{customized property}: any()
}
integrationRuntime: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'IntegrationRuntimeReference'
}
runConcurrently: any()
sourceStagingConcurrency: any()
staging: {
folderPath: any()
linkedService: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
}
traceLevel: any()
}
For ExecutePipeline, use:
type: 'ExecutePipeline'
policy: {
secureInput: bool
{customized property}: any()
}
typeProperties: {
parameters: {
{customized property}: any()
}
pipeline: {
name: 'string'
referenceName: 'string'
type: 'PipelineReference'
}
waitOnCompletion: bool
}
For ExecuteSSISPackage, use:
type: 'ExecuteSSISPackage'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
connectVia: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'IntegrationRuntimeReference'
}
environmentPath: any()
executionCredential: {
domain: any()
password: {
type: 'string'
value: 'string'
}
userName: any()
}
loggingLevel: any()
logLocation: {
logPath: any()
type: 'File'
typeProperties: {
accessCredential: {
domain: any()
password: {
type: 'string'
// For remaining properties, see SecretBase objects
}
userName: any()
}
logRefreshInterval: any()
}
}
packageConnectionManagers: {
{customized property}: {
{customized property}: {
value: any()
}
}
}
packageLocation: {
packagePath: any()
type: 'string'
typeProperties: {
accessCredential: {
domain: any()
password: {
type: 'string'
// For remaining properties, see SecretBase objects
}
userName: any()
}
childPackages: [
{
packageContent: any()
packageLastModifiedDate: 'string'
packageName: 'string'
packagePath: any()
}
]
configurationAccessCredential: {
domain: any()
password: {
type: 'string'
// For remaining properties, see SecretBase objects
}
userName: any()
}
configurationPath: any()
packageContent: any()
packageLastModifiedDate: 'string'
packageName: 'string'
packagePassword: {
type: 'string'
// For remaining properties, see SecretBase objects
}
}
}
packageParameters: {
{customized property}: {
value: any()
}
}
projectConnectionManagers: {
{customized property}: {
{customized property}: {
value: any()
}
}
}
projectParameters: {
{customized property}: {
value: any()
}
}
propertyOverrides: {
{customized property}: {
isSensitive: bool
value: any()
}
}
runtime: any()
}
For ExecuteWranglingDataflow, use:
type: 'ExecuteWranglingDataflow'
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
compute: {
computeType: any()
coreCount: any()
}
continueOnError: any()
dataFlow: {
datasetParameters: any()
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'DataFlowReference'
{customized property}: any()
}
integrationRuntime: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'IntegrationRuntimeReference'
}
queries: [
{
dataflowSinks: [
{
dataset: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'DatasetReference'
}
description: 'string'
flowlet: {
datasetParameters: any()
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'DataFlowReference'
{customized property}: any()
}
linkedService: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
name: 'string'
rejectedDataLinkedService: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
schemaLinkedService: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
script: 'string'
}
]
queryName: 'string'
}
]
runConcurrently: any()
sinks: {
{customized property}: {
dataset: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'DatasetReference'
}
description: 'string'
flowlet: {
datasetParameters: any()
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'DataFlowReference'
{customized property}: any()
}
linkedService: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
name: 'string'
rejectedDataLinkedService: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
schemaLinkedService: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
script: 'string'
}
}
sourceStagingConcurrency: any()
staging: {
folderPath: any()
linkedService: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
}
traceLevel: any()
}
For Fail, use:
type: 'Fail'
typeProperties: {
errorCode: any()
message: any()
}
For Filter, use:
type: 'Filter'
typeProperties: {
condition: {
type: 'Expression'
value: 'string'
}
items: {
type: 'Expression'
value: 'string'
}
}
For ForEach, use:
type: 'ForEach'
typeProperties: {
activities: [
{
dependsOn: [
{
activity: 'string'
dependencyConditions: [
'string'
]
{customized property}: any()
}
]
description: 'string'
name: 'string'
onInactiveMarkAs: 'string'
state: 'string'
userProperties: [
{
name: 'string'
value: any()
}
]
type: 'string'
// For remaining properties, see Activity objects
}
]
batchCount: int
isSequential: bool
items: {
type: 'Expression'
value: 'string'
}
}
For GetMetadata, use:
type: 'GetMetadata'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
dataset: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'DatasetReference'
}
fieldList: [
any
]
formatSettings: {
type: 'string'
// For remaining properties, see FormatReadSettings objects
}
storeSettings: {
disableMetricsCollection: any()
maxConcurrentConnections: any()
type: 'string'
// For remaining properties, see StoreReadSettings objects
}
}
For HDInsightHive, use:
type: 'HDInsightHive'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
arguments: [
any
]
defines: {
{customized property}: any()
}
getDebugInfo: 'string'
queryTimeout: int
scriptLinkedService: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
scriptPath: any()
storageLinkedServices: [
{
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
]
variables: {
{customized property}: any()
}
}
For HDInsightMapReduce, use:
type: 'HDInsightMapReduce'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
arguments: [
any
]
className: any()
defines: {
{customized property}: any()
}
getDebugInfo: 'string'
jarFilePath: any()
jarLibs: [
any
]
jarLinkedService: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
storageLinkedServices: [
{
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
]
}
For HDInsightPig, use:
type: 'HDInsightPig'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
arguments: any()
defines: {
{customized property}: any()
}
getDebugInfo: 'string'
scriptLinkedService: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
scriptPath: any()
storageLinkedServices: [
{
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
]
}
For HDInsightSpark, use:
type: 'HDInsightSpark'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
arguments: [
any
]
className: 'string'
entryFilePath: any()
getDebugInfo: 'string'
proxyUser: any()
rootPath: any()
sparkConfig: {
{customized property}: any()
}
sparkJobLinkedService: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
}
For HDInsightStreaming, use:
type: 'HDInsightStreaming'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
arguments: [
any
]
combiner: any()
commandEnvironment: [
any
]
defines: {
{customized property}: any()
}
fileLinkedService: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
filePaths: [
any
]
getDebugInfo: 'string'
input: any()
mapper: any()
output: any()
reducer: any()
storageLinkedServices: [
{
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
]
}
For IfCondition, use:
type: 'IfCondition'
typeProperties: {
expression: {
type: 'Expression'
value: 'string'
}
ifFalseActivities: [
{
dependsOn: [
{
activity: 'string'
dependencyConditions: [
'string'
]
{customized property}: any()
}
]
description: 'string'
name: 'string'
onInactiveMarkAs: 'string'
state: 'string'
userProperties: [
{
name: 'string'
value: any()
}
]
type: 'string'
// For remaining properties, see Activity objects
}
]
ifTrueActivities: [
{
dependsOn: [
{
activity: 'string'
dependencyConditions: [
'string'
]
{customized property}: any()
}
]
description: 'string'
name: 'string'
onInactiveMarkAs: 'string'
state: 'string'
userProperties: [
{
name: 'string'
value: any()
}
]
type: 'string'
// For remaining properties, see Activity objects
}
]
}
For Lookup, use:
type: 'Lookup'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
dataset: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'DatasetReference'
}
firstRowOnly: any()
source: {
disableMetricsCollection: any()
maxConcurrentConnections: any()
sourceRetryCount: any()
sourceRetryWait: any()
type: 'string'
// For remaining properties, see CopySource objects
}
}
For Script, use:
type: 'Script'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
logSettings: {
logDestination: 'string'
logLocationSettings: {
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
path: any()
}
}
scriptBlockExecutionTimeout: any()
scripts: [
{
parameters: [
{
direction: 'string'
name: any()
size: int
type: 'string'
value: any()
}
]
text: any()
type: 'string'
}
]
}
For SetVariable, use:
type: 'SetVariable'
policy: {
secureInput: bool
secureOutput: bool
}
typeProperties: {
setSystemVariable: bool
value: any()
variableName: 'string'
}
For SparkJob, use:
type: 'SparkJob'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
args: [
any
]
className: any()
conf: any()
configurationType: 'string'
driverSize: any()
executorSize: any()
file: any()
files: [
any
]
filesV2: [
any
]
numExecutors: any()
pythonCodeReference: [
any
]
scanFolder: any()
sparkConfig: {
{customized property}: any()
}
sparkJob: {
referenceName: any()
type: 'SparkJobDefinitionReference'
}
targetBigDataPool: {
referenceName: any()
type: 'BigDataPoolReference'
}
targetSparkConfiguration: {
referenceName: any()
type: 'SparkConfigurationReference'
}
}
For SqlServerStoredProcedure, use:
type: 'SqlServerStoredProcedure'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
storedProcedureName: any()
storedProcedureParameters: any()
}
For Switch, use:
type: 'Switch'
typeProperties: {
cases: [
{
activities: [
{
dependsOn: [
{
activity: 'string'
dependencyConditions: [
'string'
]
{customized property}: any()
}
]
description: 'string'
name: 'string'
onInactiveMarkAs: 'string'
state: 'string'
userProperties: [
{
name: 'string'
value: any()
}
]
type: 'string'
// For remaining properties, see Activity objects
}
]
value: 'string'
}
]
defaultActivities: [
{
dependsOn: [
{
activity: 'string'
dependencyConditions: [
'string'
]
{customized property}: any()
}
]
description: 'string'
name: 'string'
onInactiveMarkAs: 'string'
state: 'string'
userProperties: [
{
name: 'string'
value: any()
}
]
type: 'string'
// For remaining properties, see Activity objects
}
]
on: {
type: 'Expression'
value: 'string'
}
}
For SynapseNotebook, use:
type: 'SynapseNotebook'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
conf: any()
configurationType: 'string'
driverSize: any()
executorSize: any()
notebook: {
referenceName: any()
type: 'NotebookReference'
}
numExecutors: any()
parameters: {
{customized property}: {
type: 'string'
value: any()
}
}
sparkConfig: {
{customized property}: any()
}
sparkPool: {
referenceName: any()
type: 'BigDataPoolReference'
}
targetSparkConfiguration: {
referenceName: any()
type: 'SparkConfigurationReference'
}
}
For Until, use:
type: 'Until'
typeProperties: {
activities: [
{
dependsOn: [
{
activity: 'string'
dependencyConditions: [
'string'
]
{customized property}: any()
}
]
description: 'string'
name: 'string'
onInactiveMarkAs: 'string'
state: 'string'
userProperties: [
{
name: 'string'
value: any()
}
]
type: 'string'
// For remaining properties, see Activity objects
}
]
expression: {
type: 'Expression'
value: 'string'
}
timeout: any()
}
For Validation, use:
type: 'Validation'
typeProperties: {
childItems: any()
dataset: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'DatasetReference'
}
minimumSize: any()
sleep: any()
timeout: any()
}
For Wait, use:
type: 'Wait'
typeProperties: {
waitTimeInSeconds: any()
}
For WebActivity, use:
type: 'WebActivity'
linkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
policy: {
retry: any()
retryIntervalInSeconds: int
secureInput: bool
secureOutput: bool
timeout: any()
{customized property}: any()
}
typeProperties: {
authentication: {
credential: {
referenceName: 'string'
type: 'CredentialReference'
{customized property}: any()
}
password: {
type: 'string'
// For remaining properties, see SecretBase objects
}
pfx: {
type: 'string'
// For remaining properties, see SecretBase objects
}
resource: any()
type: 'string'
username: any()
userTenant: any()
}
body: any()
connectVia: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'IntegrationRuntimeReference'
}
datasets: [
{
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'DatasetReference'
}
]
disableCertValidation: bool
headers: {
{customized property}: 'string'
}
httpRequestTimeout: any()
linkedServices: [
{
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
]
method: 'string'
turnOffAsync: bool
url: any()
}
For WebHook, use:
type: 'WebHook'
policy: {
secureInput: bool
secureOutput: bool
}
typeProperties: {
authentication: {
credential: {
referenceName: 'string'
type: 'CredentialReference'
{customized property}: any()
}
password: {
type: 'string'
// For remaining properties, see SecretBase objects
}
pfx: {
type: 'string'
// For remaining properties, see SecretBase objects
}
resource: any()
type: 'string'
username: any()
userTenant: any()
}
body: any()
headers: {
{customized property}: 'string'
}
method: 'POST'
reportStatusOnCallBack: any()
timeout: 'string'
url: any()
}
CopySink objects
Set the type property to specify the type of object.
For AvroSink, use:
type: 'AvroSink'
formatSettings: {
fileNamePrefix: any()
maxRowsPerFile: any()
recordName: 'string'
recordNamespace: 'string'
type: 'string'
{customized property}: any()
}
storeSettings: {
copyBehavior: any()
disableMetricsCollection: any()
maxConcurrentConnections: any()
metadata: [
{
name: any()
value: any()
}
]
type: 'string'
// For remaining properties, see StoreWriteSettings objects
}
For AzureBlobFSSink, use:
type: 'AzureBlobFSSink'
copyBehavior: any()
metadata: [
{
name: any()
value: any()
}
]
For AzureDatabricksDeltaLakeSink, use:
type: 'AzureDatabricksDeltaLakeSink'
importSettings: {
dateFormat: any()
timestampFormat: any()
type: 'string'
{customized property}: any()
}
preCopyScript: any()
For AzureDataExplorerSink, use:
type: 'AzureDataExplorerSink'
flushImmediately: any()
ingestionMappingAsJson: any()
ingestionMappingName: any()
For AzureDataLakeStoreSink, use:
type: 'AzureDataLakeStoreSink'
copyBehavior: any()
enableAdlsSingleFileParallel: any()
For AzureMySqlSink, use:
type: 'AzureMySqlSink'
preCopyScript: any()
For AzurePostgreSqlSink, use:
type: 'AzurePostgreSqlSink'
preCopyScript: any()
For AzureQueueSink, use:
type: 'AzureQueueSink'
For AzureSearchIndexSink, use:
type: 'AzureSearchIndexSink'
writeBehavior: 'string'
For AzureSqlSink, use:
type: 'AzureSqlSink'
preCopyScript: any()
sqlWriterStoredProcedureName: any()
sqlWriterTableType: any()
sqlWriterUseTableLock: any()
storedProcedureParameters: any()
storedProcedureTableTypeParameterName: any()
tableOption: any()
upsertSettings: {
interimSchemaName: any()
keys: any()
useTempDB: any()
}
writeBehavior: any()
For AzureTableSink, use:
type: 'AzureTableSink'
azureTableDefaultPartitionKeyValue: any()
azureTableInsertType: any()
azureTablePartitionKeyName: any()
azureTableRowKeyName: any()
For BinarySink, use:
type: 'BinarySink'
storeSettings: {
copyBehavior: any()
disableMetricsCollection: any()
maxConcurrentConnections: any()
metadata: [
{
name: any()
value: any()
}
]
type: 'string'
// For remaining properties, see StoreWriteSettings objects
}
For BlobSink, use:
type: 'BlobSink'
blobWriterAddHeader: any()
blobWriterDateTimeFormat: any()
blobWriterOverwriteFiles: any()
copyBehavior: any()
metadata: [
{
name: any()
value: any()
}
]
For CommonDataServiceForAppsSink, use:
type: 'CommonDataServiceForAppsSink'
alternateKeyName: any()
ignoreNullValues: any()
writeBehavior: 'Upsert'
For CosmosDbMongoDbApiSink, use:
type: 'CosmosDbMongoDbApiSink'
writeBehavior: any()
For CosmosDbSqlApiSink, use:
type: 'CosmosDbSqlApiSink'
writeBehavior: any()
For DelimitedTextSink, use:
type: 'DelimitedTextSink'
formatSettings: {
fileExtension: any()
fileNamePrefix: any()
maxRowsPerFile: any()
quoteAllText: any()
type: 'string'
{customized property}: any()
}
storeSettings: {
copyBehavior: any()
disableMetricsCollection: any()
maxConcurrentConnections: any()
metadata: [
{
name: any()
value: any()
}
]
type: 'string'
// For remaining properties, see StoreWriteSettings objects
}
For DocumentDbCollectionSink, use:
type: 'DocumentDbCollectionSink'
nestingSeparator: any()
writeBehavior: any()
For DynamicsCrmSink, use:
type: 'DynamicsCrmSink'
alternateKeyName: any()
ignoreNullValues: any()
writeBehavior: 'Upsert'
For DynamicsSink, use:
type: 'DynamicsSink'
alternateKeyName: any()
ignoreNullValues: any()
writeBehavior: 'Upsert'
For FileSystemSink, use:
type: 'FileSystemSink'
copyBehavior: any()
For InformixSink, use:
type: 'InformixSink'
preCopyScript: any()
For JsonSink, use:
type: 'JsonSink'
formatSettings: {
filePattern: any()
type: 'string'
{customized property}: any()
}
storeSettings: {
copyBehavior: any()
disableMetricsCollection: any()
maxConcurrentConnections: any()
metadata: [
{
name: any()
value: any()
}
]
type: 'string'
// For remaining properties, see StoreWriteSettings objects
}
For LakeHouseTableSink, use:
type: 'LakeHouseTableSink'
partitionNameList: any()
partitionOption: any()
tableActionOption: any()
For MicrosoftAccessSink, use:
type: 'MicrosoftAccessSink'
preCopyScript: any()
For MongoDbAtlasSink, use:
type: 'MongoDbAtlasSink'
writeBehavior: any()
For MongoDbV2Sink, use:
type: 'MongoDbV2Sink'
writeBehavior: any()
For OdbcSink, use:
type: 'OdbcSink'
preCopyScript: any()
For OracleSink, use:
type: 'OracleSink'
preCopyScript: any()
For OrcSink, use:
type: 'OrcSink'
formatSettings: {
fileNamePrefix: any()
maxRowsPerFile: any()
type: 'string'
{customized property}: any()
}
storeSettings: {
copyBehavior: any()
disableMetricsCollection: any()
maxConcurrentConnections: any()
metadata: [
{
name: any()
value: any()
}
]
type: 'string'
// For remaining properties, see StoreWriteSettings objects
}
For ParquetSink, use:
type: 'ParquetSink'
formatSettings: {
fileNamePrefix: any()
maxRowsPerFile: any()
type: 'string'
{customized property}: any()
}
storeSettings: {
copyBehavior: any()
disableMetricsCollection: any()
maxConcurrentConnections: any()
metadata: [
{
name: any()
value: any()
}
]
type: 'string'
// For remaining properties, see StoreWriteSettings objects
}
For RestSink, use:
type: 'RestSink'
additionalHeaders: any()
httpCompressionType: any()
httpRequestTimeout: any()
requestInterval: any()
requestMethod: any()
For SalesforceServiceCloudSink, use:
type: 'SalesforceServiceCloudSink'
externalIdFieldName: any()
ignoreNullValues: any()
writeBehavior: 'string'
For SalesforceServiceCloudV2Sink, use:
type: 'SalesforceServiceCloudV2Sink'
externalIdFieldName: any()
ignoreNullValues: any()
writeBehavior: 'string'
For SalesforceSink, use:
type: 'SalesforceSink'
externalIdFieldName: any()
ignoreNullValues: any()
writeBehavior: 'string'
For SalesforceV2Sink, use:
type: 'SalesforceV2Sink'
externalIdFieldName: any()
ignoreNullValues: any()
writeBehavior: 'string'
For SapCloudForCustomerSink, use:
type: 'SapCloudForCustomerSink'
httpRequestTimeout: any()
writeBehavior: 'string'
For SnowflakeSink, use:
type: 'SnowflakeSink'
importSettings: {
additionalCopyOptions: {
{customized property}: any()
}
additionalFormatOptions: {
{customized property}: any()
}
type: 'string'
{customized property}: any()
}
preCopyScript: any()
For SnowflakeV2Sink, use:
type: 'SnowflakeV2Sink'
importSettings: {
additionalCopyOptions: {
{customized property}: any()
}
additionalFormatOptions: {
{customized property}: any()
}
type: 'string'
{customized property}: any()
}
preCopyScript: any()
For SqlDWSink, use:
type: 'SqlDWSink'
allowCopyCommand: any()
allowPolyBase: any()
copyCommandSettings: {
additionalOptions: {
{customized property}: 'string'
}
defaultValues: [
{
columnName: any()
defaultValue: any()
}
]
}
polyBaseSettings: {
rejectSampleValue: any()
rejectType: 'string'
rejectValue: any()
useTypeDefault: any()
{customized property}: any()
}
preCopyScript: any()
sqlWriterUseTableLock: any()
tableOption: any()
upsertSettings: {
interimSchemaName: any()
keys: any()
}
writeBehavior: any()
For SqlMISink, use:
type: 'SqlMISink'
preCopyScript: any()
sqlWriterStoredProcedureName: any()
sqlWriterTableType: any()
sqlWriterUseTableLock: any()
storedProcedureParameters: any()
storedProcedureTableTypeParameterName: any()
tableOption: any()
upsertSettings: {
interimSchemaName: any()
keys: any()
useTempDB: any()
}
writeBehavior: any()
For SqlServerSink, use:
type: 'SqlServerSink'
preCopyScript: any()
sqlWriterStoredProcedureName: any()
sqlWriterTableType: any()
sqlWriterUseTableLock: any()
storedProcedureParameters: any()
storedProcedureTableTypeParameterName: any()
tableOption: any()
upsertSettings: {
interimSchemaName: any()
keys: any()
useTempDB: any()
}
writeBehavior: any()
For SqlSink, use:
type: 'SqlSink'
preCopyScript: any()
sqlWriterStoredProcedureName: any()
sqlWriterTableType: any()
sqlWriterUseTableLock: any()
storedProcedureParameters: any()
storedProcedureTableTypeParameterName: any()
tableOption: any()
upsertSettings: {
interimSchemaName: any()
keys: any()
useTempDB: any()
}
writeBehavior: any()
For WarehouseSink, use:
type: 'WarehouseSink'
allowCopyCommand: any()
copyCommandSettings: {
additionalOptions: {
{customized property}: 'string'
}
defaultValues: [
{
columnName: any()
defaultValue: any()
}
]
}
preCopyScript: any()
tableOption: any()
writeBehavior: any()
StoreWriteSettings objects
Set the type property to specify the type of object.
For AzureBlobFSWriteSettings, use:
type: 'AzureBlobFSWriteSettings'
blockSizeInMB: any()
For AzureBlobStorageWriteSettings, use:
type: 'AzureBlobStorageWriteSettings'
blockSizeInMB: any()
For AzureDataLakeStoreWriteSettings, use:
type: 'AzureDataLakeStoreWriteSettings'
expiryDateTime: any()
For AzureFileStorageWriteSettings, use:
type: 'AzureFileStorageWriteSettings'
For FileServerWriteSettings, use:
type: 'FileServerWriteSettings'
For LakeHouseWriteSettings, use:
type: 'LakeHouseWriteSettings'
For SftpWriteSettings, use:
type: 'SftpWriteSettings'
operationTimeout: any()
useTempFileRename: any()
CopySource objects
Set the type property to specify the type of object.
For AmazonMWSSource, use:
type: 'AmazonMWSSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For AmazonRdsForOracleSource, use:
type: 'AmazonRdsForOracleSource'
additionalColumns: any()
oracleReaderQuery: any()
partitionOption: any()
partitionSettings: {
partitionColumnName: any()
partitionLowerBound: any()
partitionNames: any()
partitionUpperBound: any()
}
queryTimeout: any()
For AmazonRdsForSqlServerSource, use:
type: 'AmazonRdsForSqlServerSource'
additionalColumns: any()
isolationLevel: any()
partitionOption: any()
partitionSettings: {
partitionColumnName: any()
partitionLowerBound: any()
partitionUpperBound: any()
}
produceAdditionalTypes: any()
queryTimeout: any()
sqlReaderQuery: any()
sqlReaderStoredProcedureName: any()
storedProcedureParameters: any()
For AmazonRedshiftSource, use:
type: 'AmazonRedshiftSource'
additionalColumns: any()
query: any()
queryTimeout: any()
redshiftUnloadSettings: {
bucketName: any()
s3LinkedServiceName: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
}
For AvroSource, use:
type: 'AvroSource'
additionalColumns: any()
storeSettings: {
disableMetricsCollection: any()
maxConcurrentConnections: any()
type: 'string'
// For remaining properties, see StoreReadSettings objects
}
For AzureBlobFSSource, use:
type: 'AzureBlobFSSource'
recursive: any()
skipHeaderLineCount: any()
treatEmptyAsNull: any()
For AzureDatabricksDeltaLakeSource, use:
type: 'AzureDatabricksDeltaLakeSource'
exportSettings: {
dateFormat: any()
timestampFormat: any()
type: 'string'
{customized property}: any()
}
query: any()
For AzureDataExplorerSource, use:
type: 'AzureDataExplorerSource'
additionalColumns: any()
noTruncation: any()
query: any()
queryTimeout: any()
For AzureDataLakeStoreSource, use:
type: 'AzureDataLakeStoreSource'
recursive: any()
For AzureMariaDBSource, use:
type: 'AzureMariaDBSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For AzureMySqlSource, use:
type: 'AzureMySqlSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For AzurePostgreSqlSource, use:
type: 'AzurePostgreSqlSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For AzureSqlSource, use:
type: 'AzureSqlSource'
additionalColumns: any()
isolationLevel: any()
partitionOption: any()
partitionSettings: {
partitionColumnName: any()
partitionLowerBound: any()
partitionUpperBound: any()
}
produceAdditionalTypes: any()
queryTimeout: any()
sqlReaderQuery: any()
sqlReaderStoredProcedureName: any()
storedProcedureParameters: any()
For AzureTableSource, use:
type: 'AzureTableSource'
additionalColumns: any()
azureTableSourceIgnoreTableNotFound: any()
azureTableSourceQuery: any()
queryTimeout: any()
For BinarySource, use:
type: 'BinarySource'
formatSettings: {
compressionProperties: {
type: 'string'
// For remaining properties, see CompressionReadSettings objects
}
type: 'string'
{customized property}: any()
}
storeSettings: {
disableMetricsCollection: any()
maxConcurrentConnections: any()
type: 'string'
// For remaining properties, see StoreReadSettings objects
}
For BlobSource, use:
type: 'BlobSource'
recursive: any()
skipHeaderLineCount: any()
treatEmptyAsNull: any()
For CassandraSource, use:
type: 'CassandraSource'
additionalColumns: any()
consistencyLevel: 'string'
query: any()
queryTimeout: any()
For CommonDataServiceForAppsSource, use:
type: 'CommonDataServiceForAppsSource'
additionalColumns: any()
query: any()
For ConcurSource, use:
type: 'ConcurSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For CosmosDbMongoDbApiSource, use:
type: 'CosmosDbMongoDbApiSource'
additionalColumns: any()
batchSize: any()
cursorMethods: {
limit: any()
project: any()
skip: any()
sort: any()
{customized property}: any()
}
filter: any()
queryTimeout: any()
For CosmosDbSqlApiSource, use:
type: 'CosmosDbSqlApiSource'
additionalColumns: any()
detectDatetime: any()
pageSize: any()
preferredRegions: any()
query: any()
For CouchbaseSource, use:
type: 'CouchbaseSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For Db2Source, use:
type: 'Db2Source'
additionalColumns: any()
query: any()
queryTimeout: any()
For DelimitedTextSource, use:
type: 'DelimitedTextSource'
additionalColumns: any()
formatSettings: {
compressionProperties: {
type: 'string'
// For remaining properties, see CompressionReadSettings objects
}
skipLineCount: any()
type: 'string'
{customized property}: any()
}
storeSettings: {
disableMetricsCollection: any()
maxConcurrentConnections: any()
type: 'string'
// For remaining properties, see StoreReadSettings objects
}
For DocumentDbCollectionSource, use:
type: 'DocumentDbCollectionSource'
additionalColumns: any()
nestingSeparator: any()
query: any()
queryTimeout: any()
For DrillSource, use:
type: 'DrillSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For DynamicsAXSource, use:
type: 'DynamicsAXSource'
additionalColumns: any()
httpRequestTimeout: any()
query: any()
queryTimeout: any()
For DynamicsCrmSource, use:
type: 'DynamicsCrmSource'
additionalColumns: any()
query: any()
For DynamicsSource, use:
type: 'DynamicsSource'
additionalColumns: any()
query: any()
For EloquaSource, use:
type: 'EloquaSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For ExcelSource, use:
type: 'ExcelSource'
additionalColumns: any()
storeSettings: {
disableMetricsCollection: any()
maxConcurrentConnections: any()
type: 'string'
// For remaining properties, see StoreReadSettings objects
}
For FileSystemSource, use:
type: 'FileSystemSource'
additionalColumns: any()
recursive: any()
For GoogleAdWordsSource, use:
type: 'GoogleAdWordsSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For GoogleBigQuerySource, use:
type: 'GoogleBigQuerySource'
additionalColumns: any()
query: any()
queryTimeout: any()
For GoogleBigQueryV2Source, use:
type: 'GoogleBigQueryV2Source'
additionalColumns: any()
query: any()
queryTimeout: any()
For GreenplumSource, use:
type: 'GreenplumSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For HBaseSource, use:
type: 'HBaseSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For HdfsSource, use:
type: 'HdfsSource'
distcpSettings: {
distcpOptions: any()
resourceManagerEndpoint: any()
tempScriptPath: any()
}
recursive: any()
For HiveSource, use:
type: 'HiveSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For HttpSource, use:
type: 'HttpSource'
httpRequestTimeout: any()
For HubspotSource, use:
type: 'HubspotSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For ImpalaSource, use:
type: 'ImpalaSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For InformixSource, use:
type: 'InformixSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For JiraSource, use:
type: 'JiraSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For JsonSource, use:
type: 'JsonSource'
additionalColumns: any()
formatSettings: {
compressionProperties: {
type: 'string'
// For remaining properties, see CompressionReadSettings objects
}
type: 'string'
{customized property}: any()
}
storeSettings: {
disableMetricsCollection: any()
maxConcurrentConnections: any()
type: 'string'
// For remaining properties, see StoreReadSettings objects
}
For LakeHouseTableSource, use:
type: 'LakeHouseTableSource'
additionalColumns: any()
timestampAsOf: any()
versionAsOf: any()
For MagentoSource, use:
type: 'MagentoSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For MariaDBSource, use:
type: 'MariaDBSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For MarketoSource, use:
type: 'MarketoSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For MicrosoftAccessSource, use:
type: 'MicrosoftAccessSource'
additionalColumns: any()
query: any()
For MongoDbAtlasSource, use:
type: 'MongoDbAtlasSource'
additionalColumns: any()
batchSize: any()
cursorMethods: {
limit: any()
project: any()
skip: any()
sort: any()
{customized property}: any()
}
filter: any()
queryTimeout: any()
For MongoDbSource, use:
type: 'MongoDbSource'
additionalColumns: any()
query: any()
For MongoDbV2Source, use:
type: 'MongoDbV2Source'
additionalColumns: any()
batchSize: any()
cursorMethods: {
limit: any()
project: any()
skip: any()
sort: any()
{customized property}: any()
}
filter: any()
queryTimeout: any()
For MySqlSource, use:
type: 'MySqlSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For NetezzaSource, use:
type: 'NetezzaSource'
additionalColumns: any()
partitionOption: any()
partitionSettings: {
partitionColumnName: any()
partitionLowerBound: any()
partitionUpperBound: any()
}
query: any()
queryTimeout: any()
For ODataSource, use:
type: 'ODataSource'
additionalColumns: any()
httpRequestTimeout: any()
query: any()
For OdbcSource, use:
type: 'OdbcSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For Office365Source, use:
type: 'Office365Source'
allowedGroups: any()
dateFilterColumn: any()
endTime: any()
outputColumns: any()
startTime: any()
userScopeFilterUri: any()
For OracleServiceCloudSource, use:
type: 'OracleServiceCloudSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For OracleSource, use:
type: 'OracleSource'
additionalColumns: any()
oracleReaderQuery: any()
partitionOption: any()
partitionSettings: {
partitionColumnName: any()
partitionLowerBound: any()
partitionNames: any()
partitionUpperBound: any()
}
queryTimeout: any()
For OrcSource, use:
type: 'OrcSource'
additionalColumns: any()
storeSettings: {
disableMetricsCollection: any()
maxConcurrentConnections: any()
type: 'string'
// For remaining properties, see StoreReadSettings objects
}
For ParquetSource, use:
type: 'ParquetSource'
additionalColumns: any()
formatSettings: {
compressionProperties: {
type: 'string'
// For remaining properties, see CompressionReadSettings objects
}
type: 'string'
{customized property}: any()
}
storeSettings: {
disableMetricsCollection: any()
maxConcurrentConnections: any()
type: 'string'
// For remaining properties, see StoreReadSettings objects
}
For PaypalSource, use:
type: 'PaypalSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For PhoenixSource, use:
type: 'PhoenixSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For PostgreSqlSource, use:
type: 'PostgreSqlSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For PostgreSqlV2Source, use:
type: 'PostgreSqlV2Source'
additionalColumns: any()
query: any()
queryTimeout: any()
For PrestoSource, use:
type: 'PrestoSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For QuickBooksSource, use:
type: 'QuickBooksSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For RelationalSource, use:
type: 'RelationalSource'
additionalColumns: any()
query: any()
For ResponsysSource, use:
type: 'ResponsysSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For RestSource, use:
type: 'RestSource'
additionalColumns: any()
additionalHeaders: any()
httpRequestTimeout: any()
paginationRules: any()
requestBody: any()
requestInterval: any()
requestMethod: any()
For SalesforceMarketingCloudSource, use:
type: 'SalesforceMarketingCloudSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For SalesforceServiceCloudSource, use:
type: 'SalesforceServiceCloudSource'
additionalColumns: any()
query: any()
readBehavior: any()
For SalesforceServiceCloudV2Source, use:
type: 'SalesforceServiceCloudV2Source'
additionalColumns: any()
includeDeletedObjects: any()
SOQLQuery: any()
For SalesforceSource, use:
type: 'SalesforceSource'
additionalColumns: any()
query: any()
queryTimeout: any()
readBehavior: any()
For SalesforceV2Source, use:
type: 'SalesforceV2Source'
additionalColumns: any()
includeDeletedObjects: any()
queryTimeout: any()
SOQLQuery: any()
For SapBwSource, use:
type: 'SapBwSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For SapCloudForCustomerSource, use:
type: 'SapCloudForCustomerSource'
additionalColumns: any()
httpRequestTimeout: any()
query: any()
queryTimeout: any()
For SapEccSource, use:
type: 'SapEccSource'
additionalColumns: any()
httpRequestTimeout: any()
query: any()
queryTimeout: any()
For SapHanaSource, use:
type: 'SapHanaSource'
additionalColumns: any()
packetSize: any()
partitionOption: any()
partitionSettings: {
partitionColumnName: any()
}
query: any()
queryTimeout: any()
For SapOdpSource, use:
type: 'SapOdpSource'
additionalColumns: any()
extractionMode: any()
projection: any()
queryTimeout: any()
selection: any()
subscriberProcess: any()
For SapOpenHubSource, use:
type: 'SapOpenHubSource'
additionalColumns: any()
baseRequestId: any()
customRfcReadTableFunctionModule: any()
excludeLastRequest: any()
queryTimeout: any()
sapDataColumnDelimiter: any()
For SapTableSource, use:
type: 'SapTableSource'
additionalColumns: any()
batchSize: any()
customRfcReadTableFunctionModule: any()
partitionOption: any()
partitionSettings: {
maxPartitionsNumber: any()
partitionColumnName: any()
partitionLowerBound: any()
partitionUpperBound: any()
}
queryTimeout: any()
rfcTableFields: any()
rfcTableOptions: any()
rowCount: any()
rowSkips: any()
sapDataColumnDelimiter: any()
For ServiceNowSource, use:
type: 'ServiceNowSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For ServiceNowV2Source, use:
type: 'ServiceNowV2Source'
additionalColumns: any()
expression: {
operands: [
{
operator: 'string'
type: 'string'
value: 'string'
}
queryTimeout: any()
For SharePointOnlineListSource, use:
type: 'SharePointOnlineListSource'
httpRequestTimeout: any()
query: any()
For ShopifySource, use:
type: 'ShopifySource'
additionalColumns: any()
query: any()
queryTimeout: any()
For SnowflakeSource, use:
type: 'SnowflakeSource'
exportSettings: {
additionalCopyOptions: {
{customized property}: any()
}
additionalFormatOptions: {
{customized property}: any()
}
type: 'string'
{customized property}: any()
}
query: any()
For SnowflakeV2Source, use:
type: 'SnowflakeV2Source'
exportSettings: {
additionalCopyOptions: {
{customized property}: any()
}
additionalFormatOptions: {
{customized property}: any()
}
type: 'string'
{customized property}: any()
}
query: any()
For SparkSource, use:
type: 'SparkSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For SqlDWSource, use:
type: 'SqlDWSource'
additionalColumns: any()
isolationLevel: any()
partitionOption: any()
partitionSettings: {
partitionColumnName: any()
partitionLowerBound: any()
partitionUpperBound: any()
}
queryTimeout: any()
sqlReaderQuery: any()
sqlReaderStoredProcedureName: any()
storedProcedureParameters: any()
For SqlMISource, use:
type: 'SqlMISource'
additionalColumns: any()
isolationLevel: any()
partitionOption: any()
partitionSettings: {
partitionColumnName: any()
partitionLowerBound: any()
partitionUpperBound: any()
}
produceAdditionalTypes: any()
queryTimeout: any()
sqlReaderQuery: any()
sqlReaderStoredProcedureName: any()
storedProcedureParameters: any()
For SqlServerSource, use:
type: 'SqlServerSource'
additionalColumns: any()
isolationLevel: any()
partitionOption: any()
partitionSettings: {
partitionColumnName: any()
partitionLowerBound: any()
partitionUpperBound: any()
}
produceAdditionalTypes: any()
queryTimeout: any()
sqlReaderQuery: any()
sqlReaderStoredProcedureName: any()
storedProcedureParameters: any()
For SqlSource, use:
type: 'SqlSource'
additionalColumns: any()
isolationLevel: any()
partitionOption: any()
partitionSettings: {
partitionColumnName: any()
partitionLowerBound: any()
partitionUpperBound: any()
}
queryTimeout: any()
sqlReaderQuery: any()
sqlReaderStoredProcedureName: any()
storedProcedureParameters: any()
For SquareSource, use:
type: 'SquareSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For SybaseSource, use:
type: 'SybaseSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For TeradataSource, use:
type: 'TeradataSource'
additionalColumns: any()
partitionOption: any()
partitionSettings: {
partitionColumnName: any()
partitionLowerBound: any()
partitionUpperBound: any()
}
query: any()
queryTimeout: any()
For VerticaSource, use:
type: 'VerticaSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For WarehouseSource, use:
type: 'WarehouseSource'
additionalColumns: any()
isolationLevel: any()
partitionOption: any()
partitionSettings: {
partitionColumnName: any()
partitionLowerBound: any()
partitionUpperBound: any()
}
queryTimeout: any()
sqlReaderQuery: any()
sqlReaderStoredProcedureName: any()
storedProcedureParameters: any()
For WebSource, use:
type: 'WebSource'
additionalColumns: any()
For XeroSource, use:
type: 'XeroSource'
additionalColumns: any()
query: any()
queryTimeout: any()
For XmlSource, use:
type: 'XmlSource'
additionalColumns: any()
formatSettings: {
compressionProperties: {
type: 'string'
// For remaining properties, see CompressionReadSettings objects
}
detectDataType: any()
namespacePrefixes: any()
namespaces: any()
type: 'string'
validationMode: any()
{customized property}: any()
}
storeSettings: {
disableMetricsCollection: any()
maxConcurrentConnections: any()
type: 'string'
// For remaining properties, see StoreReadSettings objects
}
For ZohoSource, use:
type: 'ZohoSource'
additionalColumns: any()
query: any()
queryTimeout: any()
StoreReadSettings objects
Set the type property to specify the type of object.
For AmazonS3CompatibleReadSettings, use:
type: 'AmazonS3CompatibleReadSettings'
deleteFilesAfterCompletion: any()
enablePartitionDiscovery: any()
fileListPath: any()
modifiedDatetimeEnd: any()
modifiedDatetimeStart: any()
partitionRootPath: any()
prefix: any()
recursive: any()
wildcardFileName: any()
wildcardFolderPath: any()
For AmazonS3ReadSettings, use:
type: 'AmazonS3ReadSettings'
deleteFilesAfterCompletion: any()
enablePartitionDiscovery: any()
fileListPath: any()
modifiedDatetimeEnd: any()
modifiedDatetimeStart: any()
partitionRootPath: any()
prefix: any()
recursive: any()
wildcardFileName: any()
wildcardFolderPath: any()
For AzureBlobFSReadSettings, use:
type: 'AzureBlobFSReadSettings'
deleteFilesAfterCompletion: any()
enablePartitionDiscovery: any()
fileListPath: any()
modifiedDatetimeEnd: any()
modifiedDatetimeStart: any()
partitionRootPath: any()
recursive: any()
wildcardFileName: any()
wildcardFolderPath: any()
For AzureBlobStorageReadSettings, use:
type: 'AzureBlobStorageReadSettings'
deleteFilesAfterCompletion: any()
enablePartitionDiscovery: any()
fileListPath: any()
modifiedDatetimeEnd: any()
modifiedDatetimeStart: any()
partitionRootPath: any()
prefix: any()
recursive: any()
wildcardFileName: any()
wildcardFolderPath: any()
For AzureDataLakeStoreReadSettings, use:
type: 'AzureDataLakeStoreReadSettings'
deleteFilesAfterCompletion: any()
enablePartitionDiscovery: any()
fileListPath: any()
listAfter: any()
listBefore: any()
modifiedDatetimeEnd: any()
modifiedDatetimeStart: any()
partitionRootPath: any()
recursive: any()
wildcardFileName: any()
wildcardFolderPath: any()
For AzureFileStorageReadSettings, use:
type: 'AzureFileStorageReadSettings'
deleteFilesAfterCompletion: any()
enablePartitionDiscovery: any()
fileListPath: any()
modifiedDatetimeEnd: any()
modifiedDatetimeStart: any()
partitionRootPath: any()
prefix: any()
recursive: any()
wildcardFileName: any()
wildcardFolderPath: any()
For FileServerReadSettings, use:
type: 'FileServerReadSettings'
deleteFilesAfterCompletion: any()
enablePartitionDiscovery: any()
fileFilter: any()
fileListPath: any()
modifiedDatetimeEnd: any()
modifiedDatetimeStart: any()
partitionRootPath: any()
recursive: any()
wildcardFileName: any()
wildcardFolderPath: any()
For FtpReadSettings, use:
type: 'FtpReadSettings'
deleteFilesAfterCompletion: any()
disableChunking: any()
enablePartitionDiscovery: any()
fileListPath: any()
partitionRootPath: any()
recursive: any()
useBinaryTransfer: any()
wildcardFileName: any()
wildcardFolderPath: any()
For GoogleCloudStorageReadSettings, use:
type: 'GoogleCloudStorageReadSettings'
deleteFilesAfterCompletion: any()
enablePartitionDiscovery: any()
fileListPath: any()
modifiedDatetimeEnd: any()
modifiedDatetimeStart: any()
partitionRootPath: any()
prefix: any()
recursive: any()
wildcardFileName: any()
wildcardFolderPath: any()
For HdfsReadSettings, use:
type: 'HdfsReadSettings'
deleteFilesAfterCompletion: any()
distcpSettings: {
distcpOptions: any()
resourceManagerEndpoint: any()
tempScriptPath: any()
}
enablePartitionDiscovery: any()
fileListPath: any()
modifiedDatetimeEnd: any()
modifiedDatetimeStart: any()
partitionRootPath: any()
recursive: any()
wildcardFileName: any()
wildcardFolderPath: any()
For HttpReadSettings, use:
type: 'HttpReadSettings'
additionalColumns: any()
additionalHeaders: any()
requestBody: any()
requestMethod: any()
requestTimeout: any()
For LakeHouseReadSettings, use:
type: 'LakeHouseReadSettings'
deleteFilesAfterCompletion: any()
enablePartitionDiscovery: any()
fileListPath: any()
modifiedDatetimeEnd: any()
modifiedDatetimeStart: any()
partitionRootPath: any()
recursive: any()
wildcardFileName: any()
wildcardFolderPath: any()
For OracleCloudStorageReadSettings, use:
type: 'OracleCloudStorageReadSettings'
deleteFilesAfterCompletion: any()
enablePartitionDiscovery: any()
fileListPath: any()
modifiedDatetimeEnd: any()
modifiedDatetimeStart: any()
partitionRootPath: any()
prefix: any()
recursive: any()
wildcardFileName: any()
wildcardFolderPath: any()
For SftpReadSettings, use:
type: 'SftpReadSettings'
deleteFilesAfterCompletion: any()
disableChunking: any()
enablePartitionDiscovery: any()
fileListPath: any()
modifiedDatetimeEnd: any()
modifiedDatetimeStart: any()
partitionRootPath: any()
recursive: any()
wildcardFileName: any()
wildcardFolderPath: any()
CompressionReadSettings objects
Set the type property to specify the type of object.
For TarGZipReadSettings, use:
type: 'TarGZipReadSettings'
preserveCompressionFileNameAsFolder: any()
For TarReadSettings, use:
type: 'TarReadSettings'
preserveCompressionFileNameAsFolder: any()
For ZipDeflateReadSettings, use:
type: 'ZipDeflateReadSettings'
preserveZipFileNameAsFolder: any()
SecretBase objects
Set the type property to specify the type of object.
For AzureKeyVaultSecret, use:
type: 'AzureKeyVaultSecret'
secretName: any()
secretVersion: any()
store: {
parameters: {
{customized property}: any()
}
referenceName: 'string'
type: 'LinkedServiceReference'
}
For SecureString, use:
type: 'SecureString'
value: 'string'
FormatReadSettings objects
Set the type property to specify the type of object.
For BinaryReadSettings, use:
type: 'BinaryReadSettings'
compressionProperties: {
type: 'string'
// For remaining properties, see CompressionReadSettings objects
}
For DelimitedTextReadSettings, use:
type: 'DelimitedTextReadSettings'
compressionProperties: {
type: 'string'
// For remaining properties, see CompressionReadSettings objects
}
skipLineCount: any()
For JsonReadSettings, use:
type: 'JsonReadSettings'
compressionProperties: {
type: 'string'
// For remaining properties, see CompressionReadSettings objects
}
For ParquetReadSettings, use:
type: 'ParquetReadSettings'
compressionProperties: {
type: 'string'
// For remaining properties, see CompressionReadSettings objects
}
For XmlReadSettings, use:
type: 'XmlReadSettings'
compressionProperties: {
type: 'string'
// For remaining properties, see CompressionReadSettings objects
}
detectDataType: any()
namespacePrefixes: any()
namespaces: any()
validationMode: any()
Property values
factories/pipelines
Name | Description | Value |
---|---|---|
name | The resource name See how to set names and types for child resources in Bicep. |
string (required) Character limit: 1-260 Valid characters: Can't use: <>*#.%&:\\+?/ or control charactersStart with alphanumeric. |
parent | In Bicep, you can specify the parent resource for a child resource. You only need to add this property when the child resource is declared outside of the parent resource. For more information, see Child resource outside parent resource. |
Symbolic name for resource of type: factories |
properties | Properties of the pipeline. | Pipeline (required) |
Pipeline
Name | Description | Value |
---|---|---|
activities | List of activities in pipeline. | Activity[] |
annotations | List of tags that can be used for describing the Pipeline. | any[] |
concurrency | The max number of concurrent runs for the pipeline. | int Constraints: Min value = 1 |
description | The description of the pipeline. | string |
folder | The folder that this Pipeline is in. If not specified, Pipeline will appear at the root level. | PipelineFolder |
parameters | List of parameters for pipeline. | ParameterDefinitionSpecification |
policy | Pipeline Policy. | PipelinePolicy |
runDimensions | Dimensions emitted by Pipeline. | PipelineRunDimensions |
variables | List of variables for pipeline. | VariableDefinitionSpecification |
Activity
Name | Description | Value |
---|---|---|
dependsOn | Activity depends on condition. | ActivityDependency[] |
description | Activity description. | string |
name | Activity name. | string (required) |
onInactiveMarkAs | Status result of the activity when the state is set to Inactive. This is an optional property and if not provided when the activity is inactive, the status will be Succeeded by default. | 'Failed' 'Skipped' 'Succeeded' |
state | Activity state. This is an optional property and if not provided, the state will be Active by default. | 'Active' 'Inactive' |
userProperties | Activity user properties. | UserProperty[] |
type | Set the object type | AppendVariable AzureDataExplorerCommand AzureFunctionActivity AzureMLBatchExecution AzureMLExecutePipeline AzureMLUpdateResource Copy Custom DatabricksNotebook DatabricksSparkJar DatabricksSparkPython DataLakeAnalyticsU-SQL Delete ExecuteDataFlow ExecutePipeline ExecuteSSISPackage ExecuteWranglingDataflow Fail Filter ForEach GetMetadata HDInsightHive HDInsightMapReduce HDInsightPig HDInsightSpark HDInsightStreaming IfCondition Lookup Script SetVariable SparkJob SqlServerStoredProcedure Switch SynapseNotebook Until Validation Wait WebActivity WebHook (required) |
ActivityDependency
Name | Description | Value |
---|---|---|
activity | Activity name. | string (required) |
dependencyConditions | Match-Condition for the dependency. | String array containing any of: 'Completed' 'Failed' 'Skipped' 'Succeeded' (required) |
{customized property} | For Bicep, you can use the any() function. |
UserProperty
Name | Description | Value |
---|---|---|
name | User property name. | string (required) |
value | User property value. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
AppendVariableActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'AppendVariable' (required) |
typeProperties | Append Variable activity properties. | AppendVariableActivityTypeProperties (required) |
AppendVariableActivityTypeProperties
Name | Description | Value |
---|---|---|
value | Value to be appended. Type: could be a static value matching type of the variable item or Expression with resultType matching type of the variable item | For Bicep, you can use the any() function. |
variableName | Name of the variable whose value needs to be appended to. | string |
AzureDataExplorerCommandActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'AzureDataExplorerCommand' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure Data Explorer command activity properties. | AzureDataExplorerCommandActivityTypeProperties (required) |
LinkedServiceReference
Name | Description | Value |
---|---|---|
parameters | Arguments for LinkedService. | ParameterValueSpecification |
referenceName | Reference LinkedService name. | string (required) |
type | Linked service reference type. | 'LinkedServiceReference' (required) |
ParameterValueSpecification
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
ActivityPolicy
Name | Description | Value |
---|---|---|
retry | Maximum ordinary retry attempts. Default is 0. Type: integer (or Expression with resultType integer), minimum: 0. | For Bicep, you can use the any() function. |
retryIntervalInSeconds | Interval between each retry attempt (in seconds). The default is 30 sec. | int Constraints: Min value = 30 Max value = 86400 |
secureInput | When set to true, Input from activity is considered as secure and will not be logged to monitoring. | bool |
secureOutput | When set to true, Output from activity is considered as secure and will not be logged to monitoring. | bool |
timeout | Specifies the timeout for the activity to run. The default timeout is 7 days. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
{customized property} | For Bicep, you can use the any() function. |
AzureDataExplorerCommandActivityTypeProperties
Name | Description | Value |
---|---|---|
command | A control command, according to the Azure Data Explorer command syntax. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
commandTimeout | Control command timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9]))..) | For Bicep, you can use the any() function. |
AzureFunctionActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'AzureFunctionActivity' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure Function activity properties. | AzureFunctionActivityTypeProperties (required) |
AzureFunctionActivityTypeProperties
Name | Description | Value |
---|---|---|
body | Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
functionName | Name of the Function that the Azure Function Activity will call. Type: string (or Expression with resultType string) | For Bicep, you can use the any() function.(required) |
headers | Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). | AzureFunctionActivityTypePropertiesHeaders |
method | Rest API method for target endpoint. | 'DELETE' 'GET' 'HEAD' 'OPTIONS' 'POST' 'PUT' 'TRACE' (required) |
AzureFunctionActivityTypePropertiesHeaders
Name | Description | Value |
---|---|---|
{customized property} | string |
AzureMLBatchExecutionActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'AzureMLBatchExecution' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure ML Batch Execution activity properties. | AzureMLBatchExecutionActivityTypeProperties (required) |
AzureMLBatchExecutionActivityTypeProperties
Name | Description | Value |
---|---|---|
globalParameters | Key,Value pairs to be passed to the Azure ML Batch Execution Service endpoint. Keys must match the names of web service parameters defined in the published Azure ML web service. Values will be passed in the GlobalParameters property of the Azure ML batch execution request. | AzureMLBatchExecutionActivityTypePropertiesGlobalPar... |
webServiceInputs | Key,Value pairs, mapping the names of Azure ML endpoint's Web Service Inputs to AzureMLWebServiceFile objects specifying the input Blob locations.. This information will be passed in the WebServiceInputs property of the Azure ML batch execution request. | AzureMLBatchExecutionActivityTypePropertiesWebServic... |
webServiceOutputs | Key,Value pairs, mapping the names of Azure ML endpoint's Web Service Outputs to AzureMLWebServiceFile objects specifying the output Blob locations. This information will be passed in the WebServiceOutputs property of the Azure ML batch execution request. | AzureMLBatchExecutionActivityTypePropertiesWebServic... |
AzureMLBatchExecutionActivityTypePropertiesGlobalPar...
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
AzureMLBatchExecutionActivityTypePropertiesWebServic...
Name | Description | Value |
---|---|---|
{customized property} | AzureMLWebServiceFile |
AzureMLWebServiceFile
Name | Description | Value |
---|---|---|
filePath | The relative file path, including container name, in the Azure Blob Storage specified by the LinkedService. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
linkedServiceName | Reference to an Azure Storage LinkedService, where Azure ML WebService Input/Output file located. | LinkedServiceReference (required) |
AzureMLBatchExecutionActivityTypePropertiesWebServic...
Name | Description | Value |
---|---|---|
{customized property} | AzureMLWebServiceFile |
AzureMLExecutePipelineActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'AzureMLExecutePipeline' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure ML Execute Pipeline activity properties. | AzureMLExecutePipelineActivityTypeProperties (required) |
AzureMLExecutePipelineActivityTypeProperties
Name | Description | Value |
---|---|---|
continueOnStepFailure | Whether to continue execution of other steps in the PipelineRun if a step fails. This information will be passed in the continueOnStepFailure property of the published pipeline execution request. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
dataPathAssignments | Dictionary used for changing data path assignments without retraining. Values will be passed in the dataPathAssignments property of the published pipeline execution request. Type: object (or Expression with resultType object). | For Bicep, you can use the any() function. |
experimentName | Run history experiment name of the pipeline run. This information will be passed in the ExperimentName property of the published pipeline execution request. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
mlParentRunId | The parent Azure ML Service pipeline run id. This information will be passed in the ParentRunId property of the published pipeline execution request. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
mlPipelineEndpointId | ID of the published Azure ML pipeline endpoint. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
mlPipelineId | ID of the published Azure ML pipeline. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
mlPipelineParameters | Key,Value pairs to be passed to the published Azure ML pipeline endpoint. Keys must match the names of pipeline parameters defined in the published pipeline. Values will be passed in the ParameterAssignments property of the published pipeline execution request. Type: object with key value pairs (or Expression with resultType object). | For Bicep, you can use the any() function. |
version | Version of the published Azure ML pipeline endpoint. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
AzureMLUpdateResourceActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'AzureMLUpdateResource' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure ML Update Resource management activity properties. | AzureMLUpdateResourceActivityTypeProperties (required) |
AzureMLUpdateResourceActivityTypeProperties
Name | Description | Value |
---|---|---|
trainedModelFilePath | The relative file path in trainedModelLinkedService to represent the .ilearner file that will be uploaded by the update operation. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
trainedModelLinkedServiceName | Name of Azure Storage linked service holding the .ilearner file that will be uploaded by the update operation. | LinkedServiceReference (required) |
trainedModelName | Name of the Trained Model module in the Web Service experiment to be updated. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
CopyActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Copy' (required) |
inputs | List of inputs for the activity. | DatasetReference[] |
linkedServiceName | Linked service reference. | LinkedServiceReference |
outputs | List of outputs for the activity. | DatasetReference[] |
policy | Activity policy. | ActivityPolicy |
typeProperties | Copy activity properties. | CopyActivityTypeProperties (required) |
DatasetReference
Name | Description | Value |
---|---|---|
parameters | Arguments for dataset. | ParameterValueSpecification |
referenceName | Reference dataset name. | string (required) |
type | Dataset reference type. | 'DatasetReference' (required) |
CopyActivityTypeProperties
Name | Description | Value |
---|---|---|
dataIntegrationUnits | Maximum number of data integration units that can be used to perform this data movement. Type: integer (or Expression with resultType integer), minimum: 0. | For Bicep, you can use the any() function. |
enableSkipIncompatibleRow | Whether to skip incompatible row. Default value is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
enableStaging | Specifies whether to copy data via an interim staging. Default value is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
logSettings | Log settings customer needs provide when enabling log. | LogSettings |
logStorageSettings | (Deprecated. Please use LogSettings) Log storage settings customer need to provide when enabling session log. | LogStorageSettings |
parallelCopies | Maximum number of concurrent sessions opened on the source or sink to avoid overloading the data store. Type: integer (or Expression with resultType integer), minimum: 0. | For Bicep, you can use the any() function. |
preserve | Preserve rules. | any[] |
preserveRules | Preserve Rules. | any[] |
redirectIncompatibleRowSettings | Redirect incompatible row settings when EnableSkipIncompatibleRow is true. | RedirectIncompatibleRowSettings |
sink | Copy activity sink. | CopySink (required) |
skipErrorFile | Specify the fault tolerance for data consistency. | SkipErrorFile |
source | Copy activity source. | CopySource (required) |
stagingSettings | Specifies interim staging settings when EnableStaging is true. | StagingSettings |
translator | Copy activity translator. If not specified, tabular translator is used. | For Bicep, you can use the any() function. |
validateDataConsistency | Whether to enable Data Consistency validation. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
LogSettings
Name | Description | Value |
---|---|---|
copyActivityLogSettings | Specifies settings for copy activity log. | CopyActivityLogSettings |
enableCopyActivityLog | Specifies whether to enable copy activity log. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
logLocationSettings | Log location settings customer needs to provide when enabling log. | LogLocationSettings (required) |
CopyActivityLogSettings
Name | Description | Value |
---|---|---|
enableReliableLogging | Specifies whether to enable reliable logging. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
logLevel | Gets or sets the log level, support: Info, Warning. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
LogLocationSettings
Name | Description | Value |
---|---|---|
linkedServiceName | Log storage linked service reference. | LinkedServiceReference (required) |
path | The path to storage for storing detailed logs of activity execution. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
LogStorageSettings
Name | Description | Value |
---|---|---|
enableReliableLogging | Specifies whether to enable reliable logging. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
linkedServiceName | Log storage linked service reference. | LinkedServiceReference (required) |
logLevel | Gets or sets the log level, support: Info, Warning. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
path | The path to storage for storing detailed logs of activity execution. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
{customized property} | For Bicep, you can use the any() function. |
RedirectIncompatibleRowSettings
Name | Description | Value |
---|---|---|
linkedServiceName | Name of the Azure Storage, Storage SAS, or Azure Data Lake Store linked service used for redirecting incompatible row. Must be specified if redirectIncompatibleRowSettings is specified. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
path | The path for storing the redirect incompatible row data. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
{customized property} | For Bicep, you can use the any() function. |
CopySink
Name | Description | Value |
---|---|---|
disableMetricsCollection | If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
maxConcurrentConnections | The maximum concurrent connection count for the sink data store. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
sinkRetryCount | Sink retry count. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
sinkRetryWait | Sink retry wait. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
writeBatchSize | Write batch size. Type: integer (or Expression with resultType integer), minimum: 0. | For Bicep, you can use the any() function. |
writeBatchTimeout | Write batch timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
type | Set the object type | AvroSink AzureBlobFSSink AzureDatabricksDeltaLakeSink AzureDataExplorerSink AzureDataLakeStoreSink AzureMySqlSink AzurePostgreSqlSink AzureQueueSink AzureSearchIndexSink AzureSqlSink AzureTableSink BinarySink BlobSink CommonDataServiceForAppsSink CosmosDbMongoDbApiSink CosmosDbSqlApiSink DelimitedTextSink DocumentDbCollectionSink DynamicsCrmSink DynamicsSink FileSystemSink InformixSink JsonSink LakeHouseTableSink MicrosoftAccessSink MongoDbAtlasSink MongoDbV2Sink OdbcSink OracleSink OrcSink ParquetSink RestSink SalesforceServiceCloudSink SalesforceServiceCloudV2Sink SalesforceSink SalesforceV2Sink SapCloudForCustomerSink SnowflakeSink SnowflakeV2Sink SqlDWSink SqlMISink SqlServerSink SqlSink WarehouseSink (required) |
AvroSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AvroSink' (required) |
formatSettings | Avro format settings. | AvroWriteSettings |
storeSettings | Avro store settings. | StoreWriteSettings |
AvroWriteSettings
Name | Description | Value |
---|---|---|
fileNamePrefix | Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
maxRowsPerFile | Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
recordName | Top level record name in write result, which is required in AVRO spec. | string |
recordNamespace | Record namespace in the write result. | string |
type | The write setting type. | string (required) |
{customized property} | For Bicep, you can use the any() function. |
StoreWriteSettings
Name | Description | Value |
---|---|---|
copyBehavior | The type of copy behavior for copy sink. | For Bicep, you can use the any() function. |
disableMetricsCollection | If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
maxConcurrentConnections | The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
metadata | Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). | MetadataItem[] |
type | Set the object type | AzureBlobFSWriteSettings AzureBlobStorageWriteSettings AzureDataLakeStoreWriteSettings AzureFileStorageWriteSettings FileServerWriteSettings LakeHouseWriteSettings SftpWriteSettings (required) |
MetadataItem
Name | Description | Value |
---|---|---|
name | Metadata item key name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
value | Metadata item value. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
AzureBlobFSWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | 'AzureBlobFSWriteSettings' (required) |
blockSizeInMB | Indicates the block size(MB) when writing data to blob. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
AzureBlobStorageWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | 'AzureBlobStorageWriteSettings' (required) |
blockSizeInMB | Indicates the block size(MB) when writing data to blob. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
AzureDataLakeStoreWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | 'AzureDataLakeStoreWriteSettings' (required) |
expiryDateTime | Specifies the expiry time of the written files. The time is applied to the UTC time zone in the format of "2018-12-01T05:00:00Z". Default value is NULL. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
AzureFileStorageWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | 'AzureFileStorageWriteSettings' (required) |
FileServerWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | 'FileServerWriteSettings' (required) |
LakeHouseWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | 'LakeHouseWriteSettings' (required) |
SftpWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | 'SftpWriteSettings' (required) |
operationTimeout | Specifies the timeout for writing each chunk to SFTP server. Default value: 01:00:00 (one hour). Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
useTempFileRename | Upload to temporary file(s) and rename. Disable this option if your SFTP server doesn't support rename operation. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
AzureBlobFSSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureBlobFSSink' (required) |
copyBehavior | The type of copy behavior for copy sink. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
metadata | Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). | MetadataItem[] |
AzureDatabricksDeltaLakeSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureDatabricksDeltaLakeSink' (required) |
importSettings | Azure Databricks Delta Lake import settings. | AzureDatabricksDeltaLakeImportCommand |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
AzureDatabricksDeltaLakeImportCommand
Name | Description | Value |
---|---|---|
dateFormat | Specify the date format for csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
timestampFormat | Specify the timestamp format for csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
type | The import setting type. | string (required) |
{customized property} | For Bicep, you can use the any() function. |
AzureDataExplorerSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureDataExplorerSink' (required) |
flushImmediately | If set to true, any aggregation will be skipped. Default is false. Type: boolean. | For Bicep, you can use the any() function. |
ingestionMappingAsJson | An explicit column mapping description provided in a json format. Type: string. | For Bicep, you can use the any() function. |
ingestionMappingName | A name of a pre-created csv mapping that was defined on the target Kusto table. Type: string. | For Bicep, you can use the any() function. |
AzureDataLakeStoreSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureDataLakeStoreSink' (required) |
copyBehavior | The type of copy behavior for copy sink. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
enableAdlsSingleFileParallel | Single File Parallel. | For Bicep, you can use the any() function. |
AzureMySqlSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureMySqlSink' (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
AzurePostgreSqlSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzurePostgreSqlSink' (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
AzureQueueSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureQueueSink' (required) |
AzureSearchIndexSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureSearchIndexSink' (required) |
writeBehavior | Specify the write behavior when upserting documents into Azure Search Index. | 'Merge' 'Upload' |
AzureSqlSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureSqlSink' (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlWriterStoredProcedureName | SQL writer stored procedure name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlWriterTableType | SQL writer table type. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
storedProcedureParameters | SQL stored procedure parameters. | For Bicep, you can use the any() function. |
storedProcedureTableTypeParameterName | The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
upsertSettings | SQL upsert settings. | SqlUpsertSettings |
writeBehavior | Write behavior when copying data into Azure SQL. Type: SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum) | For Bicep, you can use the any() function. |
SqlUpsertSettings
Name | Description | Value |
---|---|---|
interimSchemaName | Schema name for interim table. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
keys | Key column names for unique row identification. Type: array of strings (or Expression with resultType array of strings). | For Bicep, you can use the any() function. |
useTempDB | Specifies whether to use temp db for upsert interim table. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
AzureTableSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureTableSink' (required) |
azureTableDefaultPartitionKeyValue | Azure Table default partition key value. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
azureTableInsertType | Azure Table insert type. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
azureTablePartitionKeyName | Azure Table partition key name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
azureTableRowKeyName | Azure Table row key name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
BinarySink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'BinarySink' (required) |
storeSettings | Binary store settings. | StoreWriteSettings |
BlobSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'BlobSink' (required) |
blobWriterAddHeader | Blob writer add header. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
blobWriterDateTimeFormat | Blob writer date time format. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
blobWriterOverwriteFiles | Blob writer overwrite files. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
copyBehavior | The type of copy behavior for copy sink. | For Bicep, you can use the any() function. |
metadata | Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). | MetadataItem[] |
CommonDataServiceForAppsSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'CommonDataServiceForAppsSink' (required) |
alternateKeyName | The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
ignoreNullValues | The flag indicating whether to ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
writeBehavior | The write behavior for the operation. | 'Upsert' (required) |
CosmosDbMongoDbApiSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'CosmosDbMongoDbApiSink' (required) |
writeBehavior | Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
CosmosDbSqlApiSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'CosmosDbSqlApiSink' (required) |
writeBehavior | Describes how to write data to Azure Cosmos DB. Type: string (or Expression with resultType string). Allowed values: insert and upsert. | For Bicep, you can use the any() function. |
DelimitedTextSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'DelimitedTextSink' (required) |
formatSettings | DelimitedText format settings. | DelimitedTextWriteSettings |
storeSettings | DelimitedText store settings. | StoreWriteSettings |
DelimitedTextWriteSettings
Name | Description | Value |
---|---|---|
fileExtension | The file extension used to create the files. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
fileNamePrefix | Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
maxRowsPerFile | Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
quoteAllText | Indicates whether string values should always be enclosed with quotes. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
type | The write setting type. | string (required) |
{customized property} | For Bicep, you can use the any() function. |
DocumentDbCollectionSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'DocumentDbCollectionSink' (required) |
nestingSeparator | Nested properties separator. Default is . (dot). Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
writeBehavior | Describes how to write data to Azure Cosmos DB. Type: string (or Expression with resultType string). Allowed values: insert and upsert. | For Bicep, you can use the any() function. |
DynamicsCrmSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'DynamicsCrmSink' (required) |
alternateKeyName | The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
ignoreNullValues | The flag indicating whether to ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
writeBehavior | The write behavior for the operation. | 'Upsert' (required) |
DynamicsSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'DynamicsSink' (required) |
alternateKeyName | The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
ignoreNullValues | The flag indicating whether ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
writeBehavior | The write behavior for the operation. | 'Upsert' (required) |
FileSystemSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'FileSystemSink' (required) |
copyBehavior | The type of copy behavior for copy sink. | For Bicep, you can use the any() function. |
InformixSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'InformixSink' (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
JsonSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'JsonSink' (required) |
formatSettings | Json format settings. | JsonWriteSettings |
storeSettings | Json store settings. | StoreWriteSettings |
JsonWriteSettings
Name | Description | Value |
---|---|---|
filePattern | File pattern of JSON. This setting controls the way a collection of JSON objects will be treated. The default value is 'setOfObjects'. It is case-sensitive. | For Bicep, you can use the any() function. |
type | The write setting type. | string (required) |
{customized property} | For Bicep, you can use the any() function. |
LakeHouseTableSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'LakeHouseTableSink' (required) |
partitionNameList | Specify the partition column names from sink columns. Type: array of objects (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
partitionOption | Create partitions in folder structure based on one or multiple columns. Each distinct column value (pair) will be a new partition. Possible values include: "None", "PartitionByKey". | For Bicep, you can use the any() function. |
tableActionOption | The type of table action for LakeHouse Table sink. Possible values include: "None", "Append", "Overwrite". | For Bicep, you can use the any() function. |
MicrosoftAccessSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'MicrosoftAccessSink' (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
MongoDbAtlasSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'MongoDbAtlasSink' (required) |
writeBehavior | Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
MongoDbV2Sink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'MongoDbV2Sink' (required) |
writeBehavior | Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
OdbcSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'OdbcSink' (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
OracleSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'OracleSink' (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
OrcSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'OrcSink' (required) |
formatSettings | ORC format settings. | OrcWriteSettings |
storeSettings | ORC store settings. | StoreWriteSettings |
OrcWriteSettings
Name | Description | Value |
---|---|---|
fileNamePrefix | Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
maxRowsPerFile | Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
type | The write setting type. | string (required) |
{customized property} | For Bicep, you can use the any() function. |
ParquetSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'ParquetSink' (required) |
formatSettings | Parquet format settings. | ParquetWriteSettings |
storeSettings | Parquet store settings. | StoreWriteSettings |
ParquetWriteSettings
Name | Description | Value |
---|---|---|
fileNamePrefix | Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
maxRowsPerFile | Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
type | The write setting type. | string (required) |
{customized property} | For Bicep, you can use the any() function. |
RestSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'RestSink' (required) |
additionalHeaders | The additional HTTP headers in the request to the RESTful API. Type: key value pairs (value should be string type). | For Bicep, you can use the any() function. |
httpCompressionType | Http Compression Type to Send data in compressed format with Optimal Compression Level, Default is None. And The Only Supported option is Gzip. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
requestInterval | The time to await before sending next request, in milliseconds | For Bicep, you can use the any() function. |
requestMethod | The HTTP method used to call the RESTful API. The default is POST. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SalesforceServiceCloudSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SalesforceServiceCloudSink' (required) |
externalIdFieldName | The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
ignoreNullValues | The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
writeBehavior | The write behavior for the operation. Default is Insert. | 'Insert' 'Upsert' |
SalesforceServiceCloudV2Sink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SalesforceServiceCloudV2Sink' (required) |
externalIdFieldName | The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
ignoreNullValues | The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
writeBehavior | The write behavior for the operation. Default is Insert. | 'Insert' 'Upsert' |
SalesforceSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SalesforceSink' (required) |
externalIdFieldName | The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
ignoreNullValues | The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
writeBehavior | The write behavior for the operation. Default is Insert. | 'Insert' 'Upsert' |
SalesforceV2Sink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SalesforceV2Sink' (required) |
externalIdFieldName | The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
ignoreNullValues | The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
writeBehavior | The write behavior for the operation. Default is Insert. | 'Insert' 'Upsert' |
SapCloudForCustomerSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SapCloudForCustomerSink' (required) |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
writeBehavior | The write behavior for the operation. Default is 'Insert'. | 'Insert' 'Update' |
SnowflakeSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SnowflakeSink' (required) |
importSettings | Snowflake import settings. | SnowflakeImportCopyCommand |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SnowflakeImportCopyCommand
Name | Description | Value |
---|---|---|
additionalCopyOptions | Additional copy options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": "'HH24:MI:SS.FF'" } | SnowflakeImportCopyCommandAdditionalCopyOptions |
additionalFormatOptions | Additional format options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalFormatOptions": { "FORCE": "TRUE", "LOAD_UNCERTAIN_FILES": "'FALSE'" } | SnowflakeImportCopyCommandAdditionalFormatOptions |
type | The import setting type. | string (required) |
{customized property} | For Bicep, you can use the any() function. |
SnowflakeImportCopyCommandAdditionalCopyOptions
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
SnowflakeImportCopyCommandAdditionalFormatOptions
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
SnowflakeV2Sink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SnowflakeV2Sink' (required) |
importSettings | Snowflake import settings. | SnowflakeImportCopyCommand |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SqlDWSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SqlDWSink' (required) |
allowCopyCommand | Indicates to use Copy Command to copy data into SQL Data Warehouse. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
allowPolyBase | Indicates to use PolyBase to copy data into SQL Data Warehouse when applicable. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
copyCommandSettings | Specifies Copy Command related settings when allowCopyCommand is true. | DWCopyCommandSettings |
polyBaseSettings | Specifies PolyBase-related settings when allowPolyBase is true. | PolybaseSettings |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
upsertSettings | SQL DW upsert settings. | SqlDWUpsertSettings |
writeBehavior | Write behavior when copying data into azure SQL DW. Type: SqlDWWriteBehaviorEnum (or Expression with resultType SqlDWWriteBehaviorEnum) | For Bicep, you can use the any() function. |
DWCopyCommandSettings
Name | Description | Value |
---|---|---|
additionalOptions | Additional options directly passed to SQL DW in Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalOptions": { "MAXERRORS": "1000", "DATEFORMAT": "'ymd'" } | DWCopyCommandSettingsAdditionalOptions |
defaultValues | Specifies the default values for each target column in SQL DW. The default values in the property overwrite the DEFAULT constraint set in the DB, and identity column cannot have a default value. Type: array of objects (or Expression with resultType array of objects). | DWCopyCommandDefaultValue[] |
DWCopyCommandSettingsAdditionalOptions
Name | Description | Value |
---|---|---|
{customized property} | string |
DWCopyCommandDefaultValue
Name | Description | Value |
---|---|---|
columnName | Column name. Type: object (or Expression with resultType string). | For Bicep, you can use the any() function. |
defaultValue | The default value of the column. Type: object (or Expression with resultType string). | For Bicep, you can use the any() function. |
PolybaseSettings
Name | Description | Value |
---|---|---|
rejectSampleValue | Determines the number of rows to attempt to retrieve before the PolyBase recalculates the percentage of rejected rows. Type: integer (or Expression with resultType integer), minimum: 0. | For Bicep, you can use the any() function. |
rejectType | Reject type. | 'percentage' 'value' |
rejectValue | Specifies the value or the percentage of rows that can be rejected before the query fails. Type: number (or Expression with resultType number), minimum: 0. | For Bicep, you can use the any() function. |
useTypeDefault | Specifies how to handle missing values in delimited text files when PolyBase retrieves data from the text file. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
{customized property} | For Bicep, you can use the any() function. |
SqlDWUpsertSettings
Name | Description | Value |
---|---|---|
interimSchemaName | Schema name for interim table. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
keys | Key column names for unique row identification. Type: array of strings (or Expression with resultType array of strings). | For Bicep, you can use the any() function. |
SqlMISink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SqlMISink' (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlWriterStoredProcedureName | SQL writer stored procedure name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlWriterTableType | SQL writer table type. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
storedProcedureParameters | SQL stored procedure parameters. | For Bicep, you can use the any() function. |
storedProcedureTableTypeParameterName | The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
upsertSettings | SQL upsert settings. | SqlUpsertSettings |
writeBehavior | White behavior when copying data into azure SQL MI. Type: string (or Expression with resultType string) | For Bicep, you can use the any() function. |
SqlServerSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SqlServerSink' (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlWriterStoredProcedureName | SQL writer stored procedure name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlWriterTableType | SQL writer table type. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
storedProcedureParameters | SQL stored procedure parameters. | For Bicep, you can use the any() function. |
storedProcedureTableTypeParameterName | The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
upsertSettings | SQL upsert settings. | SqlUpsertSettings |
writeBehavior | Write behavior when copying data into sql server. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SqlSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SqlSink' (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlWriterStoredProcedureName | SQL writer stored procedure name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlWriterTableType | SQL writer table type. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
storedProcedureParameters | SQL stored procedure parameters. | For Bicep, you can use the any() function. |
storedProcedureTableTypeParameterName | The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
upsertSettings | SQL upsert settings. | SqlUpsertSettings |
writeBehavior | Write behavior when copying data into sql. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
WarehouseSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'WarehouseSink' (required) |
allowCopyCommand | Indicates to use Copy Command to copy data into SQL Data Warehouse. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
copyCommandSettings | Specifies Copy Command related settings when allowCopyCommand is true. | DWCopyCommandSettings |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
writeBehavior | Write behavior when copying data into azure Microsoft Fabric Data Warehouse. Type: DWWriteBehaviorEnum (or Expression with resultType DWWriteBehaviorEnum) | For Bicep, you can use the any() function. |
SkipErrorFile
Name | Description | Value |
---|---|---|
dataInconsistency | Skip if source/sink file changed by other concurrent write. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
fileMissing | Skip if file is deleted by other client during copy. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
CopySource
AmazonMWSSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AmazonMWSSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
AmazonRdsForOracleSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AmazonRdsForOracleSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
oracleReaderQuery | AmazonRdsForOracle reader query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionOption | The partition mechanism that will be used for AmazonRdsForOracle read in parallel. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionSettings | The settings that will be leveraged for AmazonRdsForOracle source partitioning. | AmazonRdsForOraclePartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
AmazonRdsForOraclePartitionSettings
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionNames | Names of the physical partitions of AmazonRdsForOracle table. | For Bicep, you can use the any() function. |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
AmazonRdsForSqlServerSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AmazonRdsForSqlServerSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". | For Bicep, you can use the any() function. |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
produceAdditionalTypes | Which additional types to produce. | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". | For Bicep, you can use the any() function. |
SqlPartitionSettings
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column in integer or datetime type that will be used for proceeding partitioning. If not specified, the primary key of the table is auto-detected and used as the partition column. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionLowerBound | The minimum value of the partition column for partition range splitting. This value is used to decide the partition stride, not for filtering the rows in table. All rows in the table or query result will be partitioned and copied. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionUpperBound | The maximum value of the partition column for partition range splitting. This value is used to decide the partition stride, not for filtering the rows in table. All rows in the table or query result will be partitioned and copied. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
AmazonRedshiftSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AmazonRedshiftSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | Database query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
redshiftUnloadSettings | The Amazon S3 settings needed for the interim Amazon S3 when copying from Amazon Redshift with unload. With this, data from Amazon Redshift source will be unloaded into S3 first and then copied into the targeted sink from the interim S3. | RedshiftUnloadSettings |
RedshiftUnloadSettings
Name | Description | Value |
---|---|---|
bucketName | The bucket of the interim Amazon S3 which will be used to store the unloaded data from Amazon Redshift source. The bucket must be in the same region as the Amazon Redshift source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
s3LinkedServiceName | The name of the Amazon S3 linked service which will be used for the unload operation when copying from the Amazon Redshift source. | LinkedServiceReference (required) |
AvroSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AvroSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
storeSettings | Avro store settings. | StoreReadSettings |
StoreReadSettings
Name | Description | Value |
---|---|---|
disableMetricsCollection | If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
maxConcurrentConnections | The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
type | Set the object type | AmazonS3CompatibleReadSettings AmazonS3ReadSettings AzureBlobFSReadSettings AzureBlobStorageReadSettings AzureDataLakeStoreReadSettings AzureFileStorageReadSettings FileServerReadSettings FtpReadSettings GoogleCloudStorageReadSettings HdfsReadSettings HttpReadSettings LakeHouseReadSettings OracleCloudStorageReadSettings SftpReadSettings (required) |
AmazonS3CompatibleReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'AmazonS3CompatibleReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
prefix | The prefix filter for the S3 Compatible object name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
wildcardFileName | Amazon S3 Compatible wildcardFileName. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
wildcardFolderPath | Amazon S3 Compatible wildcardFolderPath. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
AmazonS3ReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'AmazonS3ReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
prefix | The prefix filter for the S3 object name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
wildcardFileName | AmazonS3 wildcardFileName. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
wildcardFolderPath | AmazonS3 wildcardFolderPath. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
AzureBlobFSReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'AzureBlobFSReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
wildcardFileName | Azure blobFS wildcardFileName. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
wildcardFolderPath | Azure blobFS wildcardFolderPath. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
AzureBlobStorageReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'AzureBlobStorageReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
prefix | The prefix filter for the Azure Blob name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
wildcardFileName | Azure blob wildcardFileName. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
wildcardFolderPath | Azure blob wildcardFolderPath. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
AzureDataLakeStoreReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'AzureDataLakeStoreReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
listAfter | Lists files after the value (exclusive) based on file/folder names’ lexicographical order. Applies under the folderPath in data set, and filter files/sub-folders under the folderPath. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
listBefore | Lists files before the value (inclusive) based on file/folder names’ lexicographical order. Applies under the folderPath in data set, and filter files/sub-folders under the folderPath. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
wildcardFileName | ADLS wildcardFileName. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
wildcardFolderPath | ADLS wildcardFolderPath. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
AzureFileStorageReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'AzureFileStorageReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
prefix | The prefix filter for the Azure File name starting from root path. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
wildcardFileName | Azure File Storage wildcardFileName. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
wildcardFolderPath | Azure File Storage wildcardFolderPath. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
FileServerReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'FileServerReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
fileFilter | Specify a filter to be used to select a subset of files in the folderPath rather than all files. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
wildcardFileName | FileServer wildcardFileName. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
wildcardFolderPath | FileServer wildcardFolderPath. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
FtpReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'FtpReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
disableChunking | If true, disable parallel reading within each file. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
useBinaryTransfer | Specify whether to use binary transfer mode for FTP stores. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
wildcardFileName | Ftp wildcardFileName. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
wildcardFolderPath | Ftp wildcardFolderPath. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
GoogleCloudStorageReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'GoogleCloudStorageReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
prefix | The prefix filter for the Google Cloud Storage object name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
wildcardFileName | Google Cloud Storage wildcardFileName. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
wildcardFolderPath | Google Cloud Storage wildcardFolderPath. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
HdfsReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'HdfsReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
distcpSettings | Specifies Distcp-related settings. | DistcpSettings |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
wildcardFileName | HDFS wildcardFileName. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
wildcardFolderPath | HDFS wildcardFolderPath. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
DistcpSettings
Name | Description | Value |
---|---|---|
distcpOptions | Specifies the Distcp options. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
resourceManagerEndpoint | Specifies the Yarn ResourceManager endpoint. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
tempScriptPath | Specifies an existing folder path which will be used to store temp Distcp command script. The script file is generated by ADF and will be removed after Copy job finished. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
HttpReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'HttpReadSettings' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
additionalHeaders | The additional HTTP headers in the request to the RESTful API. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
requestBody | The HTTP request body to the RESTful API if requestMethod is POST. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
requestMethod | The HTTP method used to call the RESTful API. The default is GET. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
requestTimeout | Specifies the timeout for a HTTP client to get HTTP response from HTTP server. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
LakeHouseReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'LakeHouseReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
wildcardFileName | Microsoft Fabric LakeHouse Files wildcardFileName. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
wildcardFolderPath | Microsoft Fabric LakeHouse Files wildcardFolderPath. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
OracleCloudStorageReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'OracleCloudStorageReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
prefix | The prefix filter for the Oracle Cloud Storage object name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
wildcardFileName | Oracle Cloud Storage wildcardFileName. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
wildcardFolderPath | Oracle Cloud Storage wildcardFolderPath. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SftpReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'SftpReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
disableChunking | If true, disable parallel reading within each file. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
wildcardFileName | Sftp wildcardFileName. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
wildcardFolderPath | Sftp wildcardFolderPath. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
AzureBlobFSSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureBlobFSSource' (required) |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
skipHeaderLineCount | Number of header lines to skip from each blob. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
treatEmptyAsNull | Treat empty as null. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
AzureDatabricksDeltaLakeSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureDatabricksDeltaLakeSource' (required) |
exportSettings | Azure Databricks Delta Lake export settings. | AzureDatabricksDeltaLakeExportCommand |
query | Azure Databricks Delta Lake Sql query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
AzureDatabricksDeltaLakeExportCommand
Name | Description | Value |
---|---|---|
dateFormat | Specify the date format for the csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
timestampFormat | Specify the timestamp format for the csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
type | The export setting type. | string (required) |
{customized property} | For Bicep, you can use the any() function. |
AzureDataExplorerSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureDataExplorerSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
noTruncation | The name of the Boolean option that controls whether truncation is applied to result-sets that go beyond a certain row-count limit. | For Bicep, you can use the any() function. |
query | Database query. Should be a Kusto Query Language (KQL) query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).. | For Bicep, you can use the any() function. |
AzureDataLakeStoreSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureDataLakeStoreSource' (required) |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
AzureMariaDBSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureMariaDBSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
AzureMySqlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureMySqlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | Database query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
AzurePostgreSqlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzurePostgreSqlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
AzureSqlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureSqlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
produceAdditionalTypes | Which additional types to produce. | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". | For Bicep, you can use the any() function. |
AzureTableSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureTableSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
azureTableSourceIgnoreTableNotFound | Azure Table source ignore table not found. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
azureTableSourceQuery | Azure Table source query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
BinarySource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'BinarySource' (required) |
formatSettings | Binary format settings. | BinaryReadSettings |
storeSettings | Binary store settings. | StoreReadSettings |
BinaryReadSettings
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
type | The read setting type. | string (required) |
{customized property} | For Bicep, you can use the any() function. |
CompressionReadSettings
Name | Description | Value |
---|---|---|
type | Set the object type | TarGZipReadSettings TarReadSettings ZipDeflateReadSettings (required) |
TarGZipReadSettings
Name | Description | Value |
---|---|---|
type | The Compression setting type. | 'TarGZipReadSettings' (required) |
preserveCompressionFileNameAsFolder | Preserve the compression file name as folder path. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
TarReadSettings
Name | Description | Value |
---|---|---|
type | The Compression setting type. | 'TarReadSettings' (required) |
preserveCompressionFileNameAsFolder | Preserve the compression file name as folder path. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
ZipDeflateReadSettings
Name | Description | Value |
---|---|---|
type | The Compression setting type. | 'ZipDeflateReadSettings' (required) |
preserveZipFileNameAsFolder | Preserve the zip file name as folder path. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
BlobSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'BlobSource' (required) |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
skipHeaderLineCount | Number of header lines to skip from each blob. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
treatEmptyAsNull | Treat empty as null. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
CassandraSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'CassandraSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
consistencyLevel | The consistency level specifies how many Cassandra servers must respond to a read request before returning data to the client application. Cassandra checks the specified number of Cassandra servers for data to satisfy the read request. Must be one of cassandraSourceReadConsistencyLevels. The default value is 'ONE'. It is case-insensitive. | 'ALL' 'EACH_QUORUM' 'LOCAL_ONE' 'LOCAL_QUORUM' 'LOCAL_SERIAL' 'ONE' 'QUORUM' 'SERIAL' 'THREE' 'TWO' |
query | Database query. Should be a SQL-92 query expression or Cassandra Query Language (CQL) command. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
CommonDataServiceForAppsSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'CommonDataServiceForAppsSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | FetchXML is a proprietary query language that is used in Microsoft Common Data Service for Apps (online & on-premises). Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
ConcurSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ConcurSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
CosmosDbMongoDbApiSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'CosmosDbMongoDbApiSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
batchSize | Specifies the number of documents to return in each batch of the response from MongoDB instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
cursorMethods | Cursor methods for Mongodb query. | MongoDbCursorMethodsProperties |
filter | Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
MongoDbCursorMethodsProperties
Name | Description | Value |
---|---|---|
limit | Specifies the maximum number of documents the server returns. limit() is analogous to the LIMIT statement in a SQL database. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
project | Specifies the fields to return in the documents that match the query filter. To return all fields in the matching documents, omit this parameter. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
skip | Specifies the how many documents skipped and where MongoDB begins returning results. This approach may be useful in implementing paginated results. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
sort | Specifies the order in which the query returns matching documents. Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
{customized property} | For Bicep, you can use the any() function. |
CosmosDbSqlApiSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'CosmosDbSqlApiSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
detectDatetime | Whether detect primitive values as datetime values. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
pageSize | Page size of the result. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
preferredRegions | Preferred regions. Type: array of strings (or Expression with resultType array of strings). | For Bicep, you can use the any() function. |
query | SQL API query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
CouchbaseSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'CouchbaseSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
Db2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | 'Db2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | Database query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
DelimitedTextSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'DelimitedTextSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
formatSettings | DelimitedText format settings. | DelimitedTextReadSettings |
storeSettings | DelimitedText store settings. | StoreReadSettings |
DelimitedTextReadSettings
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
skipLineCount | Indicates the number of non-empty rows to skip when reading data from input files. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
type | The read setting type. | string (required) |
{customized property} | For Bicep, you can use the any() function. |
DocumentDbCollectionSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'DocumentDbCollectionSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
nestingSeparator | Nested properties separator. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
query | Documents query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
DrillSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'DrillSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
DynamicsAXSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'DynamicsAXSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
DynamicsCrmSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'DynamicsCrmSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | FetchXML is a proprietary query language that is used in Microsoft Dynamics CRM (online & on-premises). Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
DynamicsSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'DynamicsSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | FetchXML is a proprietary query language that is used in Microsoft Dynamics (online & on-premises). Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
EloquaSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'EloquaSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
ExcelSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ExcelSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
storeSettings | Excel store settings. | StoreReadSettings |
FileSystemSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'FileSystemSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
GoogleAdWordsSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'GoogleAdWordsSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
GoogleBigQuerySource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'GoogleBigQuerySource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
GoogleBigQueryV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | 'GoogleBigQueryV2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
GreenplumSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'GreenplumSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
HBaseSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'HBaseSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
HdfsSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'HdfsSource' (required) |
distcpSettings | Specifies Distcp-related settings. | DistcpSettings |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
HiveSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'HiveSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
HttpSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'HttpSource' (required) |
httpRequestTimeout | Specifies the timeout for a HTTP client to get HTTP response from HTTP server. The default value is equivalent to System.Net.HttpWebRequest.Timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
HubspotSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'HubspotSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
ImpalaSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ImpalaSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
InformixSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'InformixSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | Database query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
JiraSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'JiraSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
JsonSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'JsonSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
formatSettings | Json format settings. | JsonReadSettings |
storeSettings | Json store settings. | StoreReadSettings |
JsonReadSettings
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
type | The read setting type. | string (required) |
{customized property} | For Bicep, you can use the any() function. |
LakeHouseTableSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'LakeHouseTableSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
timestampAsOf | Query an older snapshot by timestamp. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
versionAsOf | Query an older snapshot by version. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
MagentoSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MagentoSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
MariaDBSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MariaDBSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
MarketoSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MarketoSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
MicrosoftAccessSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MicrosoftAccessSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | Database query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
MongoDbAtlasSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MongoDbAtlasSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
batchSize | Specifies the number of documents to return in each batch of the response from MongoDB Atlas instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
cursorMethods | Cursor methods for Mongodb query | MongoDbCursorMethodsProperties |
filter | Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
MongoDbSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MongoDbSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | Database query. Should be a SQL-92 query expression. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
MongoDbV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MongoDbV2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
batchSize | Specifies the number of documents to return in each batch of the response from MongoDB instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
cursorMethods | Cursor methods for Mongodb query | MongoDbCursorMethodsProperties |
filter | Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
MySqlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MySqlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | Database query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
NetezzaSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'NetezzaSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
partitionOption | The partition mechanism that will be used for Netezza read in parallel. Possible values include: "None", "DataSlice", "DynamicRange". | For Bicep, you can use the any() function. |
partitionSettings | The settings that will be leveraged for Netezza source partitioning. | NetezzaPartitionSettings |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
NetezzaPartitionSettings
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
ODataSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ODataSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
query | OData query. For example, "$top=1". Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
OdbcSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'OdbcSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | Database query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
Office365Source
Name | Description | Value |
---|---|---|
type | Copy source type. | 'Office365Source' (required) |
allowedGroups | The groups containing all the users. Type: array of strings (or Expression with resultType array of strings). | For Bicep, you can use the any() function. |
dateFilterColumn | The Column to apply the {paramref name="StartTime"/} and {paramref name="EndTime"/}. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
endTime | End time of the requested range for this dataset. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
outputColumns | The columns to be read out from the Office 365 table. Type: array of objects (or Expression with resultType array of objects). itemType: OutputColumn. Example: [ { "name": "Id" }, { "name": "CreatedDateTime" } ] | For Bicep, you can use the any() function. |
startTime | Start time of the requested range for this dataset. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
userScopeFilterUri | The user scope uri. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
OracleServiceCloudSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'OracleServiceCloudSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
OracleSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'OracleSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
oracleReaderQuery | Oracle reader query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionOption | The partition mechanism that will be used for Oracle read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". | For Bicep, you can use the any() function. |
partitionSettings | The settings that will be leveraged for Oracle source partitioning. | OraclePartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
OraclePartitionSettings
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionNames | Names of the physical partitions of Oracle table. | For Bicep, you can use the any() function. |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
OrcSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'OrcSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
storeSettings | ORC store settings. | StoreReadSettings |
ParquetSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ParquetSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
formatSettings | Parquet format settings. | ParquetReadSettings |
storeSettings | Parquet store settings. | StoreReadSettings |
ParquetReadSettings
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
type | The read setting type. | string (required) |
{customized property} | For Bicep, you can use the any() function. |
PaypalSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'PaypalSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
PhoenixSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'PhoenixSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
PostgreSqlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'PostgreSqlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | Database query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
PostgreSqlV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | 'PostgreSqlV2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | Database query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
PrestoSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'PrestoSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
QuickBooksSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'QuickBooksSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
RelationalSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'RelationalSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | Database query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
ResponsysSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ResponsysSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
RestSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'RestSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: key value pairs (value should be string type). | For Bicep, you can use the any() function. |
additionalHeaders | The additional HTTP headers in the request to the RESTful API. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
paginationRules | The pagination rules to compose next page requests. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
requestBody | The HTTP request body to the RESTful API if requestMethod is POST. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
requestInterval | The time to await before sending next page request. | For Bicep, you can use the any() function. |
requestMethod | The HTTP method used to call the RESTful API. The default is GET. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SalesforceMarketingCloudSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SalesforceMarketingCloudSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
SalesforceServiceCloudSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SalesforceServiceCloudSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | Database query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
readBehavior | The read behavior for the operation. Default is Query. Allowed values: Query/QueryAll. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SalesforceServiceCloudV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SalesforceServiceCloudV2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
includeDeletedObjects | This property control whether query result contains Deleted objects. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
SOQLQuery | Database query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SalesforceSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SalesforceSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | Database query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
readBehavior | The read behavior for the operation. Default is Query. Allowed values: Query/QueryAll. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SalesforceV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SalesforceV2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
includeDeletedObjects | This property control whether query result contains Deleted objects. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
SOQLQuery | Database query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SapBwSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapBwSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | MDX query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
SapCloudForCustomerSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapCloudForCustomerSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
query | SAP Cloud for Customer OData query. For example, "$top=1". Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
SapEccSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapEccSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
query | SAP ECC OData query. For example, "$top=1". Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
SapHanaSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapHanaSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
packetSize | The packet size of data read from SAP HANA. Type: integer(or Expression with resultType integer). | For Bicep, you can use the any() function. |
partitionOption | The partition mechanism that will be used for SAP HANA read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "SapHanaDynamicRange". | For Bicep, you can use the any() function. |
partitionSettings | The settings that will be leveraged for SAP HANA source partitioning. | SapHanaPartitionSettings |
query | SAP HANA Sql query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
SapHanaPartitionSettings
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SapOdpSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapOdpSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
extractionMode | The extraction mode. Allowed value include: Full, Delta and Recovery. The default value is Full. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
projection | Specifies the columns to be selected from source data. Type: array of objects(projection) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
selection | Specifies the selection conditions from source data. Type: array of objects(selection) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
subscriberProcess | The subscriber process to manage the delta process. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SapOpenHubSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapOpenHubSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
baseRequestId | The ID of request for delta loading. Once it is set, only data with requestId larger than the value of this property will be retrieved. The default value is 0. Type: integer (or Expression with resultType integer ). | For Bicep, you can use the any() function. |
customRfcReadTableFunctionModule | Specifies the custom RFC function module that will be used to read data from SAP Table. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
excludeLastRequest | Whether to exclude the records of the last request. The default value is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
sapDataColumnDelimiter | The single character that will be used as delimiter passed to SAP RFC as well as splitting the output data retrieved. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SapTableSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapTableSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
batchSize | Specifies the maximum number of rows that will be retrieved at a time when retrieving data from SAP Table. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
customRfcReadTableFunctionModule | Specifies the custom RFC function module that will be used to read data from SAP Table. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionOption | The partition mechanism that will be used for SAP table read in parallel. Possible values include: "None", "PartitionOnInt", "PartitionOnCalendarYear", "PartitionOnCalendarMonth", "PartitionOnCalendarDate", "PartitionOnTime". | For Bicep, you can use the any() function. |
partitionSettings | The settings that will be leveraged for SAP table source partitioning. | SapTablePartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
rfcTableFields | The fields of the SAP table that will be retrieved. For example, column0, column1. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
rfcTableOptions | The options for the filtering of the SAP Table. For example, COLUMN0 EQ SOME VALUE. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
rowCount | The number of rows to be retrieved. Type: integer(or Expression with resultType integer). | For Bicep, you can use the any() function. |
rowSkips | The number of rows that will be skipped. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
sapDataColumnDelimiter | The single character that will be used as delimiter passed to SAP RFC as well as splitting the output data retrieved. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SapTablePartitionSettings
Name | Description | Value |
---|---|---|
maxPartitionsNumber | The maximum value of partitions the table will be split into. Type: integer (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionColumnName | The name of the column that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
ServiceNowSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ServiceNowSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
ServiceNowV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ServiceNowV2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
expression | Expression to filter data from source. | ExpressionV2 |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
ExpressionV2
Name | Description | Value |
---|---|---|
operands | List of nested expressions. | ExpressionV2[] |
operator | Expression operator value Type: string. | string |
type | Type of expressions supported by the system. Type: string. | 'Binary' 'Constant' 'Field' 'Unary' |
value | Value for Constant/Field Type: string. | string |
SharePointOnlineListSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SharePointOnlineListSource' (required) |
httpRequestTimeout | The wait time to get a response from SharePoint Online. Default value is 5 minutes (00:05:00). Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
query | The OData query to filter the data in SharePoint Online list. For example, "$top=1". Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
ShopifySource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ShopifySource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
SnowflakeSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SnowflakeSource' (required) |
exportSettings | Snowflake export settings. | SnowflakeExportCopyCommand (required) |
query | Snowflake Sql query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SnowflakeExportCopyCommand
Name | Description | Value |
---|---|---|
additionalCopyOptions | Additional copy options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": "'HH24:MI:SS.FF'" } | SnowflakeExportCopyCommandAdditionalCopyOptions |
additionalFormatOptions | Additional format options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalFormatOptions": { "OVERWRITE": "TRUE", "MAX_FILE_SIZE": "'FALSE'" } | SnowflakeExportCopyCommandAdditionalFormatOptions |
type | The export setting type. | string (required) |
{customized property} | For Bicep, you can use the any() function. |
SnowflakeExportCopyCommandAdditionalCopyOptions
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
SnowflakeExportCopyCommandAdditionalFormatOptions
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
SnowflakeV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SnowflakeV2Source' (required) |
exportSettings | Snowflake export settings. | SnowflakeExportCopyCommand (required) |
query | Snowflake Sql query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SparkSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SparkSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
SqlDWSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SqlDWSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
sqlReaderQuery | SQL Data Warehouse reader query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Data Warehouse source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". Type: object (or Expression with resultType object), itemType: StoredProcedureParameter. | For Bicep, you can use the any() function. |
SqlMISource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SqlMISource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
produceAdditionalTypes | Which additional types to produce. | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlReaderStoredProcedureName | Name of the stored procedure for a Azure SQL Managed Instance source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". | For Bicep, you can use the any() function. |
SqlServerSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SqlServerSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
produceAdditionalTypes | Which additional types to produce. | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". | For Bicep, you can use the any() function. |
SqlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SqlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". | For Bicep, you can use the any() function. |
SquareSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SquareSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
SybaseSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SybaseSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | Database query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
TeradataSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'TeradataSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
partitionOption | The partition mechanism that will be used for teradata read in parallel. Possible values include: "None", "Hash", "DynamicRange". | For Bicep, you can use the any() function. |
partitionSettings | The settings that will be leveraged for teradata source partitioning. | TeradataPartitionSettings |
query | Teradata query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
TeradataPartitionSettings
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column that will be used for proceeding range or hash partitioning. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
VerticaSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'VerticaSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
WarehouseSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'WarehouseSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
isolationLevel | Specifies the transaction locking behavior for the Microsoft Fabric Warehouse source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". | For Bicep, you can use the any() function. |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
sqlReaderQuery | Microsoft Fabric Warehouse reader query. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
sqlReaderStoredProcedureName | Name of the stored procedure for a Microsoft Fabric Warehouse source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". Type: object (or Expression with resultType object), itemType: StoredProcedureParameter. | For Bicep, you can use the any() function. |
WebSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'WebSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
XeroSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'XeroSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
XmlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'XmlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
formatSettings | Xml format settings. | XmlReadSettings |
storeSettings | Xml store settings. | StoreReadSettings |
XmlReadSettings
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
detectDataType | Indicates whether type detection is enabled when reading the xml files. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
namespacePrefixes | Namespace uri to prefix mappings to override the prefixes in column names when namespace is enabled, if no prefix is defined for a namespace uri, the prefix of xml element/attribute name in the xml data file will be used. Example: "{"http://www.example.com/xml":"prefix"}" Type: object (or Expression with resultType object). | For Bicep, you can use the any() function. |
namespaces | Indicates whether namespace is enabled when reading the xml files. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
type | The read setting type. | string (required) |
validationMode | Indicates what validation method is used when reading the xml files. Allowed values: 'none', 'xsd', or 'dtd'. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
{customized property} | For Bicep, you can use the any() function. |
ZohoSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ZohoSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | For Bicep, you can use the any() function. |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
StagingSettings
Name | Description | Value |
---|---|---|
enableCompression | Specifies whether to use compression when copying data via an interim staging. Default value is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
linkedServiceName | Staging linked service reference. | LinkedServiceReference (required) |
path | The path to storage for storing the interim data. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
{customized property} | For Bicep, you can use the any() function. |
CustomActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Custom' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Custom activity properties. | CustomActivityTypeProperties (required) |
CustomActivityTypeProperties
Name | Description | Value |
---|---|---|
autoUserSpecification | Elevation level and scope for the user, default is nonadmin task. Type: string (or Expression with resultType double). | For Bicep, you can use the any() function. |
command | Command for custom activity Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
extendedProperties | User defined property bag. There is no restriction on the keys or values that can be used. The user specified custom activity has the full responsibility to consume and interpret the content defined. | CustomActivityTypePropertiesExtendedProperties |
folderPath | Folder path for resource files Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
referenceObjects | Reference objects | CustomActivityReferenceObject |
resourceLinkedService | Resource linked service reference. | LinkedServiceReference |
retentionTimeInDays | The retention time for the files submitted for custom activity. Type: double (or Expression with resultType double). | For Bicep, you can use the any() function. |
CustomActivityTypePropertiesExtendedProperties
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
CustomActivityReferenceObject
Name | Description | Value |
---|---|---|
datasets | Dataset references. | DatasetReference[] |
linkedServices | Linked service references. | LinkedServiceReference[] |
DatabricksNotebookActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'DatabricksNotebook' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Databricks Notebook activity properties. | DatabricksNotebookActivityTypeProperties (required) |
DatabricksNotebookActivityTypeProperties
Name | Description | Value |
---|---|---|
baseParameters | Base parameters to be used for each run of this job.If the notebook takes a parameter that is not specified, the default value from the notebook will be used. | DatabricksNotebookActivityTypePropertiesBaseParamete... |
libraries | A list of libraries to be installed on the cluster that will execute the job. | DatabricksNotebookActivityTypePropertiesLibrariesIte...[] |
notebookPath | The absolute path of the notebook to be run in the Databricks Workspace. This path must begin with a slash. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
DatabricksNotebookActivityTypePropertiesBaseParamete...
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
DatabricksNotebookActivityTypePropertiesLibrariesIte...
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
DatabricksSparkJarActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'DatabricksSparkJar' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Databricks SparkJar activity properties. | DatabricksSparkJarActivityTypeProperties (required) |
DatabricksSparkJarActivityTypeProperties
Name | Description | Value |
---|---|---|
libraries | A list of libraries to be installed on the cluster that will execute the job. | DatabricksSparkJarActivityTypePropertiesLibrariesIte...[] |
mainClassName | The full name of the class containing the main method to be executed. This class must be contained in a JAR provided as a library. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
parameters | Parameters that will be passed to the main method. | any[] |
DatabricksSparkJarActivityTypePropertiesLibrariesIte...
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
DatabricksSparkPythonActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'DatabricksSparkPython' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Databricks SparkPython activity properties. | DatabricksSparkPythonActivityTypeProperties (required) |
DatabricksSparkPythonActivityTypeProperties
Name | Description | Value |
---|---|---|
libraries | A list of libraries to be installed on the cluster that will execute the job. | DatabricksSparkPythonActivityTypePropertiesLibraries...[] |
parameters | Command line parameters that will be passed to the Python file. | any[] |
pythonFile | The URI of the Python file to be executed. DBFS paths are supported. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
DatabricksSparkPythonActivityTypePropertiesLibraries...
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
DataLakeAnalyticsUsqlActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'DataLakeAnalyticsU-SQL' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Data Lake Analytics U-SQL activity properties. | DataLakeAnalyticsUsqlActivityTypeProperties (required) |
DataLakeAnalyticsUsqlActivityTypeProperties
Name | Description | Value |
---|---|---|
compilationMode | Compilation mode of U-SQL. Must be one of these values : Semantic, Full and SingleBox. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
degreeOfParallelism | The maximum number of nodes simultaneously used to run the job. Default value is 1. Type: integer (or Expression with resultType integer), minimum: 1. | For Bicep, you can use the any() function. |
parameters | Parameters for U-SQL job request. | DataLakeAnalyticsUsqlActivityTypePropertiesParameter... |
priority | Determines which jobs out of all that are queued should be selected to run first. The lower the number, the higher the priority. Default value is 1000. Type: integer (or Expression with resultType integer), minimum: 1. | For Bicep, you can use the any() function. |
runtimeVersion | Runtime version of the U-SQL engine to use. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
scriptLinkedService | Script linked service reference. | LinkedServiceReference (required) |
scriptPath | Case-sensitive path to folder that contains the U-SQL script. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
DataLakeAnalyticsUsqlActivityTypePropertiesParameter...
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
DeleteActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Delete' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Delete activity properties. | DeleteActivityTypeProperties (required) |
DeleteActivityTypeProperties
Name | Description | Value |
---|---|---|
dataset | Delete activity dataset reference. | DatasetReference (required) |
enableLogging | Whether to record detailed logs of delete-activity execution. Default value is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
logStorageSettings | Log storage settings customer need to provide when enableLogging is true. | LogStorageSettings |
maxConcurrentConnections | The max concurrent connections to connect data source at the same time. | int Constraints: Min value = 1 |
recursive | If true, files or sub-folders under current folder path will be deleted recursively. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
storeSettings | Delete activity store settings. | StoreReadSettings |
ExecuteDataFlowActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'ExecuteDataFlow' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute data flow activity properties. | ExecuteDataFlowActivityTypeProperties (required) |
ExecuteDataFlowActivityTypeProperties
Name | Description | Value |
---|---|---|
compute | Compute properties for data flow activity. | ExecuteDataFlowActivityTypePropertiesCompute |
continueOnError | Continue on error setting used for data flow execution. Enables processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean) | For Bicep, you can use the any() function. |
dataFlow | Data flow reference. | DataFlowReference (required) |
integrationRuntime | The integration runtime reference. | IntegrationRuntimeReference |
runConcurrently | Concurrent run setting used for data flow execution. Allows sinks with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean) | For Bicep, you can use the any() function. |
sourceStagingConcurrency | Specify number of parallel staging for sources applicable to the sink. Type: integer (or Expression with resultType integer) | For Bicep, you can use the any() function. |
staging | Staging info for execute data flow activity. | DataFlowStagingInfo |
traceLevel | Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string) | For Bicep, you can use the any() function. |
ExecuteDataFlowActivityTypePropertiesCompute
Name | Description | Value |
---|---|---|
computeType | Compute type of the cluster which will execute data flow job. Possible values include: 'General', 'MemoryOptimized', 'ComputeOptimized'. Type: string (or Expression with resultType string) | For Bicep, you can use the any() function. |
coreCount | Core count of the cluster which will execute data flow job. Supported values are: 8, 16, 32, 48, 80, 144 and 272. Type: integer (or Expression with resultType integer) | For Bicep, you can use the any() function. |
DataFlowReference
Name | Description | Value |
---|---|---|
datasetParameters | Reference data flow parameters from dataset. | For Bicep, you can use the any() function. |
parameters | Data flow parameters | ParameterValueSpecification |
referenceName | Reference data flow name. | string (required) |
type | Data flow reference type. | 'DataFlowReference' (required) |
{customized property} | For Bicep, you can use the any() function. |
IntegrationRuntimeReference
Name | Description | Value |
---|---|---|
parameters | Arguments for integration runtime. | ParameterValueSpecification |
referenceName | Reference integration runtime name. | string (required) |
type | Type of integration runtime. | 'IntegrationRuntimeReference' (required) |
DataFlowStagingInfo
Name | Description | Value |
---|---|---|
folderPath | Folder path for staging blob. Type: string (or Expression with resultType string) | For Bicep, you can use the any() function. |
linkedService | Staging linked service reference. | LinkedServiceReference |
ExecutePipelineActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'ExecutePipeline' (required) |
policy | Execute pipeline activity policy. | ExecutePipelineActivityPolicy |
typeProperties | Execute pipeline activity properties. | ExecutePipelineActivityTypeProperties (required) |
ExecutePipelineActivityPolicy
Name | Description | Value |
---|---|---|
secureInput | When set to true, Input from activity is considered as secure and will not be logged to monitoring. | bool |
{customized property} | For Bicep, you can use the any() function. |
ExecutePipelineActivityTypeProperties
Name | Description | Value |
---|---|---|
parameters | Pipeline parameters. | ParameterValueSpecification |
pipeline | Pipeline reference. | PipelineReference (required) |
waitOnCompletion | Defines whether activity execution will wait for the dependent pipeline execution to finish. Default is false. | bool |
PipelineReference
Name | Description | Value |
---|---|---|
name | Reference name. | string |
referenceName | Reference pipeline name. | string (required) |
type | Pipeline reference type. | 'PipelineReference' (required) |
ExecuteSsisPackageActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'ExecuteSSISPackage' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute SSIS package activity properties. | ExecuteSsisPackageActivityTypeProperties (required) |
ExecuteSsisPackageActivityTypeProperties
Name | Description | Value |
---|---|---|
connectVia | The integration runtime reference. | IntegrationRuntimeReference (required) |
environmentPath | The environment path to execute the SSIS package. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
executionCredential | The package execution credential. | SsisExecutionCredential |
loggingLevel | The logging level of SSIS package execution. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
logLocation | SSIS package execution log location. | SsisLogLocation |
packageConnectionManagers | The package level connection managers to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesPackageConne... |
packageLocation | SSIS package location. | SsisPackageLocation (required) |
packageParameters | The package level parameters to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesPackageParam... |
projectConnectionManagers | The project level connection managers to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesProjectConne... |
projectParameters | The project level parameters to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesProjectParam... |
propertyOverrides | The property overrides to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesPropertyOver... |
runtime | Specifies the runtime to execute SSIS package. The value should be "x86" or "x64". Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SsisExecutionCredential
Name | Description | Value |
---|---|---|
domain | Domain for windows authentication. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
password | Password for windows authentication. | SecureString (required) |
userName | UseName for windows authentication. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
SecureString
Name | Description | Value |
---|---|---|
type | Type of the secret. | string (required) |
value | Value of secure string. | string (required) |
SsisLogLocation
Name | Description | Value |
---|---|---|
logPath | The SSIS package execution log path. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
type | The type of SSIS log location. | 'File' (required) |
typeProperties | SSIS package execution log location properties. | SsisLogLocationTypeProperties (required) |
SsisLogLocationTypeProperties
Name | Description | Value |
---|---|---|
accessCredential | The package execution log access credential. | SsisAccessCredential |
logRefreshInterval | Specifies the interval to refresh log. The default interval is 5 minutes. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
SsisAccessCredential
Name | Description | Value |
---|---|---|
domain | Domain for windows authentication. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
password | Password for windows authentication. | SecretBase (required) |
userName | UseName for windows authentication. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
SecretBase
Name | Description | Value |
---|---|---|
type | Set the object type | AzureKeyVaultSecret SecureString (required) |
AzureKeyVaultSecretReference
Name | Description | Value |
---|---|---|
secretName | The name of the secret in Azure Key Vault. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
secretVersion | The version of the secret in Azure Key Vault. The default value is the latest version of the secret. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
store | The Azure Key Vault linked service reference. | LinkedServiceReference (required) |
type | Type of the secret. | string (required) |
ExecuteSsisPackageActivityTypePropertiesPackageConne...
Name | Description | Value |
---|---|---|
{customized property} | SsisConnectionManager |
SsisConnectionManager
Name | Description | Value |
---|---|---|
{customized property} | SsisExecutionParameter |
SsisExecutionParameter
Name | Description | Value |
---|---|---|
value | SSIS package execution parameter value. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
SsisPackageLocation
Name | Description | Value |
---|---|---|
packagePath | The SSIS package path. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
type | The type of SSIS package location. | 'File' 'InlinePackage' 'PackageStore' 'SSISDB' |
typeProperties | SSIS package location properties. | SsisPackageLocationTypeProperties |
SsisPackageLocationTypeProperties
Name | Description | Value |
---|---|---|
accessCredential | The package access credential. | SsisAccessCredential |
childPackages | The embedded child package list. | SsisChildPackage[] |
configurationAccessCredential | The configuration file access credential. | SsisAccessCredential |
configurationPath | The configuration file of the package execution. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
packageContent | The embedded package content. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
packageLastModifiedDate | The embedded package last modified date. | string |
packageName | The package name. | string |
packagePassword | Password of the package. | SecretBase |
SsisChildPackage
Name | Description | Value |
---|---|---|
packageContent | Content for embedded child package. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
packageLastModifiedDate | Last modified date for embedded child package. | string |
packageName | Name for embedded child package. | string |
packagePath | Path for embedded child package. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
ExecuteSsisPackageActivityTypePropertiesPackageParam...
Name | Description | Value |
---|---|---|
{customized property} | SsisExecutionParameter |
ExecuteSsisPackageActivityTypePropertiesProjectConne...
Name | Description | Value |
---|---|---|
{customized property} | SsisConnectionManager |
ExecuteSsisPackageActivityTypePropertiesProjectParam...
Name | Description | Value |
---|---|---|
{customized property} | SsisExecutionParameter |
ExecuteSsisPackageActivityTypePropertiesPropertyOver...
Name | Description | Value |
---|---|---|
{customized property} | SsisPropertyOverride |
SsisPropertyOverride
Name | Description | Value |
---|---|---|
isSensitive | Whether SSIS package property override value is sensitive data. Value will be encrypted in SSISDB if it is true | bool |
value | SSIS package property override value. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
ExecuteWranglingDataflowActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'ExecuteWranglingDataflow' (required) |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute power query activity properties. | ExecutePowerQueryActivityTypeProperties (required) |
ExecutePowerQueryActivityTypeProperties
Name | Description | Value |
---|---|---|
compute | Compute properties for data flow activity. | ExecuteDataFlowActivityTypePropertiesCompute |
continueOnError | Continue on error setting used for data flow execution. Enables processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean) | For Bicep, you can use the any() function. |
dataFlow | Data flow reference. | DataFlowReference (required) |
integrationRuntime | The integration runtime reference. | IntegrationRuntimeReference |
queries | List of mapping for Power Query mashup query to sink dataset(s). | PowerQuerySinkMapping[] |
runConcurrently | Concurrent run setting used for data flow execution. Allows sinks with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean) | For Bicep, you can use the any() function. |
sinks | (Deprecated. Please use Queries). List of Power Query activity sinks mapped to a queryName. | ExecutePowerQueryActivityTypePropertiesSinks |
sourceStagingConcurrency | Specify number of parallel staging for sources applicable to the sink. Type: integer (or Expression with resultType integer) | For Bicep, you can use the any() function. |
staging | Staging info for execute data flow activity. | DataFlowStagingInfo |
traceLevel | Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string) | For Bicep, you can use the any() function. |
PowerQuerySinkMapping
Name | Description | Value |
---|---|---|
dataflowSinks | List of sinks mapped to Power Query mashup query. | PowerQuerySink[] |
queryName | Name of the query in Power Query mashup document. | string |
PowerQuerySink
Name | Description | Value |
---|---|---|
dataset | Dataset reference. | DatasetReference |
description | Transformation description. | string |
flowlet | Flowlet Reference | DataFlowReference |
linkedService | Linked service reference. | LinkedServiceReference |
name | Transformation name. | string (required) |
rejectedDataLinkedService | Rejected data linked service reference. | LinkedServiceReference |
schemaLinkedService | Schema linked service reference. | LinkedServiceReference |
script | sink script. | string |
ExecutePowerQueryActivityTypePropertiesSinks
Name | Description | Value |
---|---|---|
{customized property} | PowerQuerySink |
FailActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Fail' (required) |
typeProperties | Fail activity properties. | FailActivityTypeProperties (required) |
FailActivityTypeProperties
Name | Description | Value |
---|---|---|
errorCode | The error code that categorizes the error type of the Fail activity. It can be dynamic content that's evaluated to a non empty/blank string at runtime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
message | The error message that surfaced in the Fail activity. It can be dynamic content that's evaluated to a non empty/blank string at runtime. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
FilterActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Filter' (required) |
typeProperties | Filter activity properties. | FilterActivityTypeProperties (required) |
FilterActivityTypeProperties
Name | Description | Value |
---|---|---|
condition | Condition to be used for filtering the input. | Expression (required) |
items | Input array on which filter should be applied. | Expression (required) |
Expression
Name | Description | Value |
---|---|---|
type | Expression type. | 'Expression' (required) |
value | Expression value. | string (required) |
ForEachActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'ForEach' (required) |
typeProperties | ForEach activity properties. | ForEachActivityTypeProperties (required) |
ForEachActivityTypeProperties
Name | Description | Value |
---|---|---|
activities | List of activities to execute . | Activity[] (required) |
batchCount | Batch count to be used for controlling the number of parallel execution (when isSequential is set to false). | int Constraints: Max value = 50 |
isSequential | Should the loop be executed in sequence or in parallel (max 50) | bool |
items | Collection to iterate. | Expression (required) |
GetMetadataActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'GetMetadata' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | GetMetadata activity properties. | GetMetadataActivityTypeProperties (required) |
GetMetadataActivityTypeProperties
Name | Description | Value |
---|---|---|
dataset | GetMetadata activity dataset reference. | DatasetReference (required) |
fieldList | Fields of metadata to get from dataset. | any[] |
formatSettings | GetMetadata activity format settings. | FormatReadSettings |
storeSettings | GetMetadata activity store settings. | StoreReadSettings |
FormatReadSettings
Name | Description | Value |
---|---|---|
type | Set the object type | BinaryReadSettings DelimitedTextReadSettings JsonReadSettings ParquetReadSettings XmlReadSettings (required) |
HDInsightHiveActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'HDInsightHive' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight Hive activity properties. | HDInsightHiveActivityTypeProperties (required) |
HDInsightHiveActivityTypeProperties
Name | Description | Value |
---|---|---|
arguments | User specified arguments to HDInsightActivity. | any[] |
defines | Allows user to specify defines for Hive job request. | HDInsightHiveActivityTypePropertiesDefines |
getDebugInfo | Debug info option. | 'Always' 'Failure' 'None' |
queryTimeout | Query timeout value (in minutes). Effective when the HDInsight cluster is with ESP (Enterprise Security Package) | int |
scriptLinkedService | Script linked service reference. | LinkedServiceReference |
scriptPath | Script path. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
storageLinkedServices | Storage linked service references. | LinkedServiceReference[] |
variables | User specified arguments under hivevar namespace. | HDInsightHiveActivityTypePropertiesVariables |
HDInsightHiveActivityTypePropertiesDefines
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
HDInsightHiveActivityTypePropertiesVariables
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
HDInsightMapReduceActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'HDInsightMapReduce' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight MapReduce activity properties. | HDInsightMapReduceActivityTypeProperties (required) |
HDInsightMapReduceActivityTypeProperties
Name | Description | Value |
---|---|---|
arguments | User specified arguments to HDInsightActivity. | any[] |
className | Class name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
defines | Allows user to specify defines for the MapReduce job request. | HDInsightMapReduceActivityTypePropertiesDefines |
getDebugInfo | Debug info option. | 'Always' 'Failure' 'None' |
jarFilePath | Jar path. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
jarLibs | Jar libs. | any[] |
jarLinkedService | Jar linked service reference. | LinkedServiceReference |
storageLinkedServices | Storage linked service references. | LinkedServiceReference[] |
HDInsightMapReduceActivityTypePropertiesDefines
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
HDInsightPigActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'HDInsightPig' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight Pig activity properties. | HDInsightPigActivityTypeProperties (required) |
HDInsightPigActivityTypeProperties
Name | Description | Value |
---|---|---|
arguments | User specified arguments to HDInsightActivity. Type: array (or Expression with resultType array). | For Bicep, you can use the any() function. |
defines | Allows user to specify defines for Pig job request. | HDInsightPigActivityTypePropertiesDefines |
getDebugInfo | Debug info option. | 'Always' 'Failure' 'None' |
scriptLinkedService | Script linked service reference. | LinkedServiceReference |
scriptPath | Script path. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
storageLinkedServices | Storage linked service references. | LinkedServiceReference[] |
HDInsightPigActivityTypePropertiesDefines
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
HDInsightSparkActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'HDInsightSpark' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight spark activity properties. | HDInsightSparkActivityTypeProperties (required) |
HDInsightSparkActivityTypeProperties
Name | Description | Value |
---|---|---|
arguments | The user-specified arguments to HDInsightSparkActivity. | any[] |
className | The application's Java/Spark main class. | string |
entryFilePath | The relative path to the root folder of the code/package to be executed. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
getDebugInfo | Debug info option. | 'Always' 'Failure' 'None' |
proxyUser | The user to impersonate that will execute the job. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
rootPath | The root path in 'sparkJobLinkedService' for all the job’s files. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
sparkConfig | Spark configuration property. | HDInsightSparkActivityTypePropertiesSparkConfig |
sparkJobLinkedService | The storage linked service for uploading the entry file and dependencies, and for receiving logs. | LinkedServiceReference |
HDInsightSparkActivityTypePropertiesSparkConfig
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
HDInsightStreamingActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'HDInsightStreaming' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight streaming activity properties. | HDInsightStreamingActivityTypeProperties (required) |
HDInsightStreamingActivityTypeProperties
Name | Description | Value |
---|---|---|
arguments | User specified arguments to HDInsightActivity. | any[] |
combiner | Combiner executable name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
commandEnvironment | Command line environment values. | any[] |
defines | Allows user to specify defines for streaming job request. | HDInsightStreamingActivityTypePropertiesDefines |
fileLinkedService | Linked service reference where the files are located. | LinkedServiceReference |
filePaths | Paths to streaming job files. Can be directories. | any[] (required) |
getDebugInfo | Debug info option. | 'Always' 'Failure' 'None' |
input | Input blob path. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
mapper | Mapper executable name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
output | Output blob path. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
reducer | Reducer executable name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
storageLinkedServices | Storage linked service references. | LinkedServiceReference[] |
HDInsightStreamingActivityTypePropertiesDefines
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
IfConditionActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'IfCondition' (required) |
typeProperties | IfCondition activity properties. | IfConditionActivityTypeProperties (required) |
IfConditionActivityTypeProperties
Name | Description | Value |
---|---|---|
expression | An expression that would evaluate to Boolean. This is used to determine the block of activities (ifTrueActivities or ifFalseActivities) that will be executed. | Expression (required) |
ifFalseActivities | List of activities to execute if expression is evaluated to false. This is an optional property and if not provided, the activity will exit without any action. | Activity[] |
ifTrueActivities | List of activities to execute if expression is evaluated to true. This is an optional property and if not provided, the activity will exit without any action. | Activity[] |
LookupActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Lookup' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Lookup activity properties. | LookupActivityTypeProperties (required) |
LookupActivityTypeProperties
Name | Description | Value |
---|---|---|
dataset | Lookup activity dataset reference. | DatasetReference (required) |
firstRowOnly | Whether to return first row or all rows. Default value is true. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
source | Dataset-specific source properties, same as copy activity source. | CopySource (required) |
ScriptActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Script' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Script activity properties. | ScriptActivityTypeProperties (required) |
ScriptActivityTypeProperties
Name | Description | Value |
---|---|---|
logSettings | Log settings of script activity. | ScriptActivityTypePropertiesLogSettings |
scriptBlockExecutionTimeout | ScriptBlock execution timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
scripts | Array of script blocks. Type: array. | ScriptActivityScriptBlock[] |
ScriptActivityTypePropertiesLogSettings
Name | Description | Value |
---|---|---|
logDestination | The destination of logs. Type: string. | 'ActivityOutput' 'ExternalStore' (required) |
logLocationSettings | Log location settings customer needs to provide when enabling log. | LogLocationSettings |
ScriptActivityScriptBlock
Name | Description | Value |
---|---|---|
parameters | Array of script parameters. Type: array. | ScriptActivityParameter[] |
text | The query text. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
type | The type of the query. Type: string. | 'NonQuery' 'Query' (required) |
ScriptActivityParameter
Name | Description | Value |
---|---|---|
direction | The direction of the parameter. | 'Input' 'InputOutput' 'Output' |
name | The name of the parameter. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
size | The size of the output direction parameter. | int |
type | The type of the parameter. | 'Boolean' 'DateTime' 'DateTimeOffset' 'Decimal' 'Double' 'Guid' 'Int16' 'Int32' 'Int64' 'Single' 'String' 'Timespan' |
value | The value of the parameter. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SetVariableActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'SetVariable' (required) |
policy | Activity policy. | SecureInputOutputPolicy |
typeProperties | Set Variable activity properties. | SetVariableActivityTypeProperties (required) |
SecureInputOutputPolicy
Name | Description | Value |
---|---|---|
secureInput | When set to true, Input from activity is considered as secure and will not be logged to monitoring. | bool |
secureOutput | When set to true, Output from activity is considered as secure and will not be logged to monitoring. | bool |
SetVariableActivityTypeProperties
Name | Description | Value |
---|---|---|
setSystemVariable | If set to true, it sets the pipeline run return value. | bool |
value | Value to be set. Could be a static value or Expression. | For Bicep, you can use the any() function. |
variableName | Name of the variable whose value needs to be set. | string |
SynapseSparkJobDefinitionActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'SparkJob' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute spark job activity properties. | SynapseSparkJobActivityTypeProperties (required) |
SynapseSparkJobActivityTypeProperties
Name | Description | Value |
---|---|---|
args | User specified arguments to SynapseSparkJobDefinitionActivity. | any[] |
className | The fully-qualified identifier or the main class that is in the main definition file, which will override the 'className' of the spark job definition you provide. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
conf | Spark configuration properties, which will override the 'conf' of the spark job definition you provide. | For Bicep, you can use the any() function. |
configurationType | The type of the spark config. | 'Artifact' 'Customized' 'Default' |
driverSize | Number of core and memory to be used for driver allocated in the specified Spark pool for the job, which will be used for overriding 'driverCores' and 'driverMemory' of the spark job definition you provide. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
executorSize | Number of core and memory to be used for executors allocated in the specified Spark pool for the job, which will be used for overriding 'executorCores' and 'executorMemory' of the spark job definition you provide. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
file | The main file used for the job, which will override the 'file' of the spark job definition you provide. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
files | (Deprecated. Please use pythonCodeReference and filesV2) Additional files used for reference in the main definition file, which will override the 'files' of the spark job definition you provide. | any[] |
filesV2 | Additional files used for reference in the main definition file, which will override the 'jars' and 'files' of the spark job definition you provide. | any[] |
numExecutors | Number of executors to launch for this job, which will override the 'numExecutors' of the spark job definition you provide. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
pythonCodeReference | Additional python code files used for reference in the main definition file, which will override the 'pyFiles' of the spark job definition you provide. | any[] |
scanFolder | Scanning subfolders from the root folder of the main definition file, these files will be added as reference files. The folders named 'jars', 'pyFiles', 'files' or 'archives' will be scanned, and the folders name are case sensitive. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
sparkConfig | Spark configuration property. | SynapseSparkJobActivityTypePropertiesSparkConfig |
sparkJob | Synapse spark job reference. | SynapseSparkJobReference (required) |
targetBigDataPool | The name of the big data pool which will be used to execute the spark batch job, which will override the 'targetBigDataPool' of the spark job definition you provide. | BigDataPoolParametrizationReference |
targetSparkConfiguration | The spark configuration of the spark job. | SparkConfigurationParametrizationReference |
SynapseSparkJobActivityTypePropertiesSparkConfig
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
SynapseSparkJobReference
Name | Description | Value |
---|---|---|
referenceName | Reference spark job name. Expression with resultType string. | For Bicep, you can use the any() function.(required) |
type | Synapse spark job reference type. | 'SparkJobDefinitionReference' (required) |
BigDataPoolParametrizationReference
Name | Description | Value |
---|---|---|
referenceName | Reference big data pool name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
type | Big data pool reference type. | 'BigDataPoolReference' (required) |
SparkConfigurationParametrizationReference
Name | Description | Value |
---|---|---|
referenceName | Reference spark configuration name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
type | Spark configuration reference type. | 'SparkConfigurationReference' (required) |
SqlServerStoredProcedureActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'SqlServerStoredProcedure' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | SQL stored procedure activity properties. | SqlServerStoredProcedureActivityTypeProperties (required) |
SqlServerStoredProcedureActivityTypeProperties
Name | Description | Value |
---|---|---|
storedProcedureName | Stored procedure name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". | For Bicep, you can use the any() function. |
SwitchActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Switch' (required) |
typeProperties | Switch activity properties. | SwitchActivityTypeProperties (required) |
SwitchActivityTypeProperties
Name | Description | Value |
---|---|---|
cases | List of cases that correspond to expected values of the 'on' property. This is an optional property and if not provided, the activity will execute activities provided in defaultActivities. | SwitchCase[] |
defaultActivities | List of activities to execute if no case condition is satisfied. This is an optional property and if not provided, the activity will exit without any action. | Activity[] |
on | An expression that would evaluate to a string or integer. This is used to determine the block of activities in cases that will be executed. | Expression (required) |
SwitchCase
Name | Description | Value |
---|---|---|
activities | List of activities to execute for satisfied case condition. | Activity[] |
value | Expected value that satisfies the expression result of the 'on' property. | string |
SynapseNotebookActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'SynapseNotebook' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute Synapse notebook activity properties. | SynapseNotebookActivityTypeProperties (required) |
SynapseNotebookActivityTypeProperties
Name | Description | Value |
---|---|---|
conf | Spark configuration properties, which will override the 'conf' of the notebook you provide. | For Bicep, you can use the any() function. |
configurationType | The type of the spark config. | 'Artifact' 'Customized' 'Default' |
driverSize | Number of core and memory to be used for driver allocated in the specified Spark pool for the session, which will be used for overriding 'driverCores' and 'driverMemory' of the notebook you provide. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
executorSize | Number of core and memory to be used for executors allocated in the specified Spark pool for the session, which will be used for overriding 'executorCores' and 'executorMemory' of the notebook you provide. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
notebook | Synapse notebook reference. | SynapseNotebookReference (required) |
numExecutors | Number of executors to launch for this session, which will override the 'numExecutors' of the notebook you provide. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
parameters | Notebook parameters. | SynapseNotebookActivityTypePropertiesParameters |
sparkConfig | Spark configuration property. | SynapseNotebookActivityTypePropertiesSparkConfig |
sparkPool | The name of the big data pool which will be used to execute the notebook. | BigDataPoolParametrizationReference |
targetSparkConfiguration | The spark configuration of the spark job. | SparkConfigurationParametrizationReference |
SynapseNotebookReference
Name | Description | Value |
---|---|---|
referenceName | Reference notebook name. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
type | Synapse notebook reference type. | 'NotebookReference' (required) |
SynapseNotebookActivityTypePropertiesParameters
Name | Description | Value |
---|---|---|
{customized property} | NotebookParameter |
NotebookParameter
Name | Description | Value |
---|---|---|
type | Notebook parameter type. | 'bool' 'float' 'int' 'string' |
value | Notebook parameter value. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
SynapseNotebookActivityTypePropertiesSparkConfig
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
UntilActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Until' (required) |
typeProperties | Until activity properties. | UntilActivityTypeProperties (required) |
UntilActivityTypeProperties
Name | Description | Value |
---|---|---|
activities | List of activities to execute. | Activity[] (required) |
expression | An expression that would evaluate to Boolean. The loop will continue until this expression evaluates to true | Expression (required) |
timeout | Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
ValidationActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Validation' (required) |
typeProperties | Validation activity properties. | ValidationActivityTypeProperties (required) |
ValidationActivityTypeProperties
Name | Description | Value |
---|---|---|
childItems | Can be used if dataset points to a folder. If set to true, the folder must have at least one file. If set to false, the folder must be empty. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
dataset | Validation activity dataset reference. | DatasetReference (required) |
minimumSize | Can be used if dataset points to a file. The file must be greater than or equal in size to the value specified. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
sleep | A delay in seconds between validation attempts. If no value is specified, 10 seconds will be used as the default. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function. |
timeout | Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | For Bicep, you can use the any() function. |
WaitActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Wait' (required) |
typeProperties | Wait activity properties. | WaitActivityTypeProperties (required) |
WaitActivityTypeProperties
Name | Description | Value |
---|---|---|
waitTimeInSeconds | Duration in seconds. Type: integer (or Expression with resultType integer). | For Bicep, you can use the any() function.(required) |
WebActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'WebActivity' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Web activity properties. | WebActivityTypeProperties (required) |
WebActivityTypeProperties
Name | Description | Value |
---|---|---|
authentication | Authentication method used for calling the endpoint. | WebActivityAuthentication |
body | Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
connectVia | The integration runtime reference. | IntegrationRuntimeReference |
datasets | List of datasets passed to web endpoint. | DatasetReference[] |
disableCertValidation | When set to true, Certificate validation will be disabled. | bool |
headers | Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). | WebActivityTypePropertiesHeaders |
httpRequestTimeout | Timeout for the HTTP request to get a response. Format is in TimeSpan (hh:mm:ss). This value is the timeout to get a response, not the activity timeout. The default value is 00:01:00 (1 minute). The range is from 1 to 10 minutes | For Bicep, you can use the any() function. |
linkedServices | List of linked services passed to web endpoint. | LinkedServiceReference[] |
method | Rest API method for target endpoint. | 'DELETE' 'GET' 'POST' 'PUT' (required) |
turnOffAsync | Option to disable invoking HTTP GET on location given in response header of a HTTP 202 Response. If set true, it stops invoking HTTP GET on http location given in response header. If set false then continues to invoke HTTP GET call on location given in http response headers. | bool |
url | Web activity target endpoint and path. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
WebActivityAuthentication
Name | Description | Value |
---|---|---|
credential | The credential reference containing authentication information. | CredentialReference |
password | Password for the PFX file or basic authentication / Secret when used for ServicePrincipal | SecretBase |
pfx | Base64-encoded contents of a PFX file or Certificate when used for ServicePrincipal | SecretBase |
resource | Resource for which Azure Auth token will be requested when using MSI Authentication. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
type | Web activity authentication (Basic/ClientCertificate/MSI/ServicePrincipal) | string |
username | Web activity authentication user name for basic authentication or ClientID when used for ServicePrincipal. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
userTenant | TenantId for which Azure Auth token will be requested when using ServicePrincipal Authentication. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
CredentialReference
Name | Description | Value |
---|---|---|
referenceName | Reference credential name. | string (required) |
type | Credential reference type. | 'CredentialReference' (required) |
{customized property} | For Bicep, you can use the any() function. |
WebActivityTypePropertiesHeaders
Name | Description | Value |
---|---|---|
{customized property} | string |
WebHookActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'WebHook' (required) |
policy | Activity policy. | SecureInputOutputPolicy |
typeProperties | WebHook activity properties. | WebHookActivityTypeProperties (required) |
WebHookActivityTypeProperties
Name | Description | Value |
---|---|---|
authentication | Authentication method used for calling the endpoint. | WebActivityAuthentication |
body | Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). | For Bicep, you can use the any() function. |
headers | Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). | WebHookActivityTypePropertiesHeaders |
method | Rest API method for target endpoint. | 'POST' (required) |
reportStatusOnCallBack | When set to true, statusCode, output and error in callback request body will be consumed by activity. The activity can be marked as failed by setting statusCode >= 400 in callback request. Default is false. Type: boolean (or Expression with resultType boolean). | For Bicep, you can use the any() function. |
timeout | The timeout within which the webhook should be called back. If there is no value specified, it defaults to 10 minutes. Type: string. Pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | string |
url | WebHook activity target endpoint and path. Type: string (or Expression with resultType string). | For Bicep, you can use the any() function.(required) |
WebHookActivityTypePropertiesHeaders
Name | Description | Value |
---|---|---|
{customized property} | string |
PipelineFolder
Name | Description | Value |
---|---|---|
name | The name of the folder that this Pipeline is in. | string |
ParameterDefinitionSpecification
Name | Description | Value |
---|---|---|
{customized property} | ParameterSpecification |
ParameterSpecification
Name | Description | Value |
---|---|---|
defaultValue | Default value of parameter. | For Bicep, you can use the any() function. |
type | Parameter type. | 'Array' 'Bool' 'Float' 'Int' 'Object' 'SecureString' 'String' (required) |
PipelinePolicy
Name | Description | Value |
---|---|---|
elapsedTimeMetric | Pipeline ElapsedTime Metric Policy. | PipelineElapsedTimeMetricPolicy |
PipelineElapsedTimeMetricPolicy
Name | Description | Value |
---|---|---|
duration | TimeSpan value, after which an Azure Monitoring Metric is fired. | For Bicep, you can use the any() function. |
PipelineRunDimensions
Name | Description | Value |
---|---|---|
{customized property} | For Bicep, you can use the any() function. |
VariableDefinitionSpecification
Name | Description | Value |
---|---|---|
{customized property} | VariableSpecification |
VariableSpecification
Name | Description | Value |
---|---|---|
defaultValue | Default value of variable. | For Bicep, you can use the any() function. |
type | Variable type. | 'Array' 'Bool' 'String' (required) |
Quickstart templates
The following quickstart templates deploy this resource type.
Template | Description |
---|---|
More is possible with Azure Data Factory - One click to try Azure Data Factory |
This template creates a data factory pipeline for a copy activity from Azure Blob into another Azure Blob |
Create a Data Factory, copy from Salesforce to Blob |
This template uses Azure Data Factory to perform data movement from Salesforce to Azure Blobs |
Bulk copy using Azure Data Factory |
This template creates a V2 data factory that copies data in bulk an Azure SQL database to an Azure SQL data warehouse. |
Create a V2 data factory |
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to another folder in the storage. |
Create a V2 data factory (MySQL) |
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to a table in an Azure Database for MySQL. |
Create a V2 data factory (PostGre) |
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to a table in an Azure Database for PostgreSQL. |
Create a V2 data factory (SQL) |
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to a table in an Azure SQL database. |
Create a V2 data factory (SQL On-prem) |
This template creates a V2 data factory that copies data from an on-premises SQL Server to an Azure blob storage |
Create a V2 data factory (Spark) |
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to another folder in the storage. |
ARM template resource definition
The factories/pipelines resource type can be deployed with operations that target:
- Resource groups - See resource group deployment commands
For a list of changed properties in each API version, see change log.
Resource format
To create a Microsoft.DataFactory/factories/pipelines resource, add the following JSON to your template.
{
"type": "Microsoft.DataFactory/factories/pipelines",
"apiVersion": "2018-06-01",
"name": "string",
"properties": {
"activities": [
{
"dependsOn": [
{
"activity": "string",
"dependencyConditions": [ "string" ],
"{customized property}": {}
}
],
"description": "string",
"name": "string",
"onInactiveMarkAs": "string",
"state": "string",
"userProperties": [
{
"name": "string",
"value": {}
}
],
"type": "string"
// For remaining properties, see Activity objects
}
],
"annotations": [ object ],
"concurrency": "int",
"description": "string",
"folder": {
"name": "string"
},
"parameters": {
"{customized property}": {
"defaultValue": {},
"type": "string"
}
},
"policy": {
"elapsedTimeMetric": {
"duration": {}
}
},
"runDimensions": {
"{customized property}": {}
},
"variables": {
"{customized property}": {
"defaultValue": {},
"type": "string"
}
}
}
}
Activity objects
Set the type property to specify the type of object.
For AppendVariable, use:
"type": "AppendVariable",
"typeProperties": {
"value": {},
"variableName": "string"
}
For AzureDataExplorerCommand, use:
"type": "AzureDataExplorerCommand",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"command": {},
"commandTimeout": {}
}
For AzureFunctionActivity, use:
"type": "AzureFunctionActivity",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"body": {},
"functionName": {},
"headers": {
"{customized property}": "string"
},
"method": "string"
}
For AzureMLBatchExecution, use:
"type": "AzureMLBatchExecution",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"globalParameters": {
"{customized property}": {}
},
"webServiceInputs": {
"{customized property}": {
"filePath": {},
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
}
},
"webServiceOutputs": {
"{customized property}": {
"filePath": {},
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
}
}
}
For AzureMLExecutePipeline, use:
"type": "AzureMLExecutePipeline",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"continueOnStepFailure": {},
"dataPathAssignments": {},
"experimentName": {},
"mlParentRunId": {},
"mlPipelineEndpointId": {},
"mlPipelineId": {},
"mlPipelineParameters": {},
"version": {}
}
For AzureMLUpdateResource, use:
"type": "AzureMLUpdateResource",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"trainedModelFilePath": {},
"trainedModelLinkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"trainedModelName": {}
}
For Copy, use:
"type": "Copy",
"inputs": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
}
],
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"outputs": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
}
],
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"dataIntegrationUnits": {},
"enableSkipIncompatibleRow": {},
"enableStaging": {},
"logSettings": {
"copyActivityLogSettings": {
"enableReliableLogging": {},
"logLevel": {}
},
"enableCopyActivityLog": {},
"logLocationSettings": {
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"path": {}
}
},
"logStorageSettings": {
"enableReliableLogging": {},
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"logLevel": {},
"path": {},
"{customized property}": {}
},
"parallelCopies": {},
"preserve": [ object ],
"preserveRules": [ object ],
"redirectIncompatibleRowSettings": {
"linkedServiceName": {},
"path": {},
"{customized property}": {}
},
"sink": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"sinkRetryCount": {},
"sinkRetryWait": {},
"writeBatchSize": {},
"writeBatchTimeout": {},
"type": "string"
// For remaining properties, see CopySink objects
},
"skipErrorFile": {
"dataInconsistency": {},
"fileMissing": {}
},
"source": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"sourceRetryCount": {},
"sourceRetryWait": {},
"type": "string"
// For remaining properties, see CopySource objects
},
"stagingSettings": {
"enableCompression": {},
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"path": {},
"{customized property}": {}
},
"translator": {},
"validateDataConsistency": {}
}
For Custom, use:
"type": "Custom",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"autoUserSpecification": {},
"command": {},
"extendedProperties": {
"{customized property}": {}
},
"folderPath": {},
"referenceObjects": {
"datasets": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
}
],
"linkedServices": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
]
},
"resourceLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"retentionTimeInDays": {}
}
For DatabricksNotebook, use:
"type": "DatabricksNotebook",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"baseParameters": {
"{customized property}": {}
},
"libraries": [
{
"{customized property}": {}
}
],
"notebookPath": {}
}
For DatabricksSparkJar, use:
"type": "DatabricksSparkJar",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"libraries": [
{
"{customized property}": {}
}
],
"mainClassName": {},
"parameters": [ object ]
}
For DatabricksSparkPython, use:
"type": "DatabricksSparkPython",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"libraries": [
{
"{customized property}": {}
}
],
"parameters": [ object ],
"pythonFile": {}
}
For DataLakeAnalyticsU-SQL, use:
"type": "DataLakeAnalyticsU-SQL",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"compilationMode": {},
"degreeOfParallelism": {},
"parameters": {
"{customized property}": {}
},
"priority": {},
"runtimeVersion": {},
"scriptLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"scriptPath": {}
}
For Delete, use:
"type": "Delete",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"dataset": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
},
"enableLogging": {},
"logStorageSettings": {
"enableReliableLogging": {},
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"logLevel": {},
"path": {},
"{customized property}": {}
},
"maxConcurrentConnections": "int",
"recursive": {},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
}
For ExecuteDataFlow, use:
"type": "ExecuteDataFlow",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"compute": {
"computeType": {},
"coreCount": {}
},
"continueOnError": {},
"dataFlow": {
"datasetParameters": {},
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DataFlowReference",
"{customized property}": {}
},
"integrationRuntime": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "IntegrationRuntimeReference"
},
"runConcurrently": {},
"sourceStagingConcurrency": {},
"staging": {
"folderPath": {},
"linkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
},
"traceLevel": {}
}
For ExecutePipeline, use:
"type": "ExecutePipeline",
"policy": {
"secureInput": "bool",
"{customized property}": {}
},
"typeProperties": {
"parameters": {
"{customized property}": {}
},
"pipeline": {
"name": "string",
"referenceName": "string",
"type": "PipelineReference"
},
"waitOnCompletion": "bool"
}
For ExecuteSSISPackage, use:
"type": "ExecuteSSISPackage",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"connectVia": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "IntegrationRuntimeReference"
},
"environmentPath": {},
"executionCredential": {
"domain": {},
"password": {
"type": "string",
"value": "string"
},
"userName": {}
},
"loggingLevel": {},
"logLocation": {
"logPath": {},
"type": "File",
"typeProperties": {
"accessCredential": {
"domain": {},
"password": {
"type": "string"
// For remaining properties, see SecretBase objects
},
"userName": {}
},
"logRefreshInterval": {}
}
},
"packageConnectionManagers": {
"{customized property}": {
"{customized property}": {
"value": {}
}
}
},
"packageLocation": {
"packagePath": {},
"type": "string",
"typeProperties": {
"accessCredential": {
"domain": {},
"password": {
"type": "string"
// For remaining properties, see SecretBase objects
},
"userName": {}
},
"childPackages": [
{
"packageContent": {},
"packageLastModifiedDate": "string",
"packageName": "string",
"packagePath": {}
}
],
"configurationAccessCredential": {
"domain": {},
"password": {
"type": "string"
// For remaining properties, see SecretBase objects
},
"userName": {}
},
"configurationPath": {},
"packageContent": {},
"packageLastModifiedDate": "string",
"packageName": "string",
"packagePassword": {
"type": "string"
// For remaining properties, see SecretBase objects
}
}
},
"packageParameters": {
"{customized property}": {
"value": {}
}
},
"projectConnectionManagers": {
"{customized property}": {
"{customized property}": {
"value": {}
}
}
},
"projectParameters": {
"{customized property}": {
"value": {}
}
},
"propertyOverrides": {
"{customized property}": {
"isSensitive": "bool",
"value": {}
}
},
"runtime": {}
}
For ExecuteWranglingDataflow, use:
"type": "ExecuteWranglingDataflow",
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"compute": {
"computeType": {},
"coreCount": {}
},
"continueOnError": {},
"dataFlow": {
"datasetParameters": {},
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DataFlowReference",
"{customized property}": {}
},
"integrationRuntime": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "IntegrationRuntimeReference"
},
"queries": [
{
"dataflowSinks": [
{
"dataset": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
},
"description": "string",
"flowlet": {
"datasetParameters": {},
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DataFlowReference",
"{customized property}": {}
},
"linkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"name": "string",
"rejectedDataLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"schemaLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"script": "string"
}
],
"queryName": "string"
}
],
"runConcurrently": {},
"sinks": {
"{customized property}": {
"dataset": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
},
"description": "string",
"flowlet": {
"datasetParameters": {},
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DataFlowReference",
"{customized property}": {}
},
"linkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"name": "string",
"rejectedDataLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"schemaLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"script": "string"
}
},
"sourceStagingConcurrency": {},
"staging": {
"folderPath": {},
"linkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
},
"traceLevel": {}
}
For Fail, use:
"type": "Fail",
"typeProperties": {
"errorCode": {},
"message": {}
}
For Filter, use:
"type": "Filter",
"typeProperties": {
"condition": {
"type": "Expression",
"value": "string"
},
"items": {
"type": "Expression",
"value": "string"
}
}
For ForEach, use:
"type": "ForEach",
"typeProperties": {
"activities": [
{
"dependsOn": [
{
"activity": "string",
"dependencyConditions": [ "string" ],
"{customized property}": {}
}
],
"description": "string",
"name": "string",
"onInactiveMarkAs": "string",
"state": "string",
"userProperties": [
{
"name": "string",
"value": {}
}
],
"type": "string"
// For remaining properties, see Activity objects
}
],
"batchCount": "int",
"isSequential": "bool",
"items": {
"type": "Expression",
"value": "string"
}
}
For GetMetadata, use:
"type": "GetMetadata",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"dataset": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
},
"fieldList": [ object ],
"formatSettings": {
"type": "string"
// For remaining properties, see FormatReadSettings objects
},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
}
For HDInsightHive, use:
"type": "HDInsightHive",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"arguments": [ object ],
"defines": {
"{customized property}": {}
},
"getDebugInfo": "string",
"queryTimeout": "int",
"scriptLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"scriptPath": {},
"storageLinkedServices": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
],
"variables": {
"{customized property}": {}
}
}
For HDInsightMapReduce, use:
"type": "HDInsightMapReduce",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"arguments": [ object ],
"className": {},
"defines": {
"{customized property}": {}
},
"getDebugInfo": "string",
"jarFilePath": {},
"jarLibs": [ object ],
"jarLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"storageLinkedServices": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
]
}
For HDInsightPig, use:
"type": "HDInsightPig",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"arguments": {},
"defines": {
"{customized property}": {}
},
"getDebugInfo": "string",
"scriptLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"scriptPath": {},
"storageLinkedServices": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
]
}
For HDInsightSpark, use:
"type": "HDInsightSpark",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"arguments": [ object ],
"className": "string",
"entryFilePath": {},
"getDebugInfo": "string",
"proxyUser": {},
"rootPath": {},
"sparkConfig": {
"{customized property}": {}
},
"sparkJobLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
}
For HDInsightStreaming, use:
"type": "HDInsightStreaming",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"arguments": [ object ],
"combiner": {},
"commandEnvironment": [ object ],
"defines": {
"{customized property}": {}
},
"fileLinkedService": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"filePaths": [ object ],
"getDebugInfo": "string",
"input": {},
"mapper": {},
"output": {},
"reducer": {},
"storageLinkedServices": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
]
}
For IfCondition, use:
"type": "IfCondition",
"typeProperties": {
"expression": {
"type": "Expression",
"value": "string"
},
"ifFalseActivities": [
{
"dependsOn": [
{
"activity": "string",
"dependencyConditions": [ "string" ],
"{customized property}": {}
}
],
"description": "string",
"name": "string",
"onInactiveMarkAs": "string",
"state": "string",
"userProperties": [
{
"name": "string",
"value": {}
}
],
"type": "string"
// For remaining properties, see Activity objects
}
],
"ifTrueActivities": [
{
"dependsOn": [
{
"activity": "string",
"dependencyConditions": [ "string" ],
"{customized property}": {}
}
],
"description": "string",
"name": "string",
"onInactiveMarkAs": "string",
"state": "string",
"userProperties": [
{
"name": "string",
"value": {}
}
],
"type": "string"
// For remaining properties, see Activity objects
}
]
}
For Lookup, use:
"type": "Lookup",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"dataset": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
},
"firstRowOnly": {},
"source": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"sourceRetryCount": {},
"sourceRetryWait": {},
"type": "string"
// For remaining properties, see CopySource objects
}
}
For Script, use:
"type": "Script",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"logSettings": {
"logDestination": "string",
"logLocationSettings": {
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"path": {}
}
},
"scriptBlockExecutionTimeout": {},
"scripts": [
{
"parameters": [
{
"direction": "string",
"name": {},
"size": "int",
"type": "string",
"value": {}
}
],
"text": {},
"type": "string"
}
]
}
For SetVariable, use:
"type": "SetVariable",
"policy": {
"secureInput": "bool",
"secureOutput": "bool"
},
"typeProperties": {
"setSystemVariable": "bool",
"value": {},
"variableName": "string"
}
For SparkJob, use:
"type": "SparkJob",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"args": [ object ],
"className": {},
"conf": {},
"configurationType": "string",
"driverSize": {},
"executorSize": {},
"file": {},
"files": [ object ],
"filesV2": [ object ],
"numExecutors": {},
"pythonCodeReference": [ object ],
"scanFolder": {},
"sparkConfig": {
"{customized property}": {}
},
"sparkJob": {
"referenceName": {},
"type": "SparkJobDefinitionReference"
},
"targetBigDataPool": {
"referenceName": {},
"type": "BigDataPoolReference"
},
"targetSparkConfiguration": {
"referenceName": {},
"type": "SparkConfigurationReference"
}
}
For SqlServerStoredProcedure, use:
"type": "SqlServerStoredProcedure",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"storedProcedureName": {},
"storedProcedureParameters": {}
}
For Switch, use:
"type": "Switch",
"typeProperties": {
"cases": [
{
"activities": [
{
"dependsOn": [
{
"activity": "string",
"dependencyConditions": [ "string" ],
"{customized property}": {}
}
],
"description": "string",
"name": "string",
"onInactiveMarkAs": "string",
"state": "string",
"userProperties": [
{
"name": "string",
"value": {}
}
],
"type": "string"
// For remaining properties, see Activity objects
}
],
"value": "string"
}
],
"defaultActivities": [
{
"dependsOn": [
{
"activity": "string",
"dependencyConditions": [ "string" ],
"{customized property}": {}
}
],
"description": "string",
"name": "string",
"onInactiveMarkAs": "string",
"state": "string",
"userProperties": [
{
"name": "string",
"value": {}
}
],
"type": "string"
// For remaining properties, see Activity objects
}
],
"on": {
"type": "Expression",
"value": "string"
}
}
For SynapseNotebook, use:
"type": "SynapseNotebook",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"conf": {},
"configurationType": "string",
"driverSize": {},
"executorSize": {},
"notebook": {
"referenceName": {},
"type": "NotebookReference"
},
"numExecutors": {},
"parameters": {
"{customized property}": {
"type": "string",
"value": {}
}
},
"sparkConfig": {
"{customized property}": {}
},
"sparkPool": {
"referenceName": {},
"type": "BigDataPoolReference"
},
"targetSparkConfiguration": {
"referenceName": {},
"type": "SparkConfigurationReference"
}
}
For Until, use:
"type": "Until",
"typeProperties": {
"activities": [
{
"dependsOn": [
{
"activity": "string",
"dependencyConditions": [ "string" ],
"{customized property}": {}
}
],
"description": "string",
"name": "string",
"onInactiveMarkAs": "string",
"state": "string",
"userProperties": [
{
"name": "string",
"value": {}
}
],
"type": "string"
// For remaining properties, see Activity objects
}
],
"expression": {
"type": "Expression",
"value": "string"
},
"timeout": {}
}
For Validation, use:
"type": "Validation",
"typeProperties": {
"childItems": {},
"dataset": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
},
"minimumSize": {},
"sleep": {},
"timeout": {}
}
For Wait, use:
"type": "Wait",
"typeProperties": {
"waitTimeInSeconds": {}
}
For WebActivity, use:
"type": "WebActivity",
"linkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
},
"policy": {
"retry": {},
"retryIntervalInSeconds": "int",
"secureInput": "bool",
"secureOutput": "bool",
"timeout": {},
"{customized property}": {}
},
"typeProperties": {
"authentication": {
"credential": {
"referenceName": "string",
"type": "CredentialReference",
"{customized property}": {}
},
"password": {
"type": "string"
// For remaining properties, see SecretBase objects
},
"pfx": {
"type": "string"
// For remaining properties, see SecretBase objects
},
"resource": {},
"type": "string",
"username": {},
"userTenant": {}
},
"body": {},
"connectVia": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "IntegrationRuntimeReference"
},
"datasets": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "DatasetReference"
}
],
"disableCertValidation": "bool",
"headers": {
"{customized property}": "string"
},
"httpRequestTimeout": {},
"linkedServices": [
{
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
],
"method": "string",
"turnOffAsync": "bool",
"url": {}
}
For WebHook, use:
"type": "WebHook",
"policy": {
"secureInput": "bool",
"secureOutput": "bool"
},
"typeProperties": {
"authentication": {
"credential": {
"referenceName": "string",
"type": "CredentialReference",
"{customized property}": {}
},
"password": {
"type": "string"
// For remaining properties, see SecretBase objects
},
"pfx": {
"type": "string"
// For remaining properties, see SecretBase objects
},
"resource": {},
"type": "string",
"username": {},
"userTenant": {}
},
"body": {},
"headers": {
"{customized property}": "string"
},
"method": "POST",
"reportStatusOnCallBack": {},
"timeout": "string",
"url": {}
}
CopySink objects
Set the type property to specify the type of object.
For AvroSink, use:
"type": "AvroSink",
"formatSettings": {
"fileNamePrefix": {},
"maxRowsPerFile": {},
"recordName": "string",
"recordNamespace": "string",
"type": "string",
"{customized property}": {}
},
"storeSettings": {
"copyBehavior": {},
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"metadata": [
{
"name": {},
"value": {}
}
],
"type": "string"
// For remaining properties, see StoreWriteSettings objects
}
For AzureBlobFSSink, use:
"type": "AzureBlobFSSink",
"copyBehavior": {},
"metadata": [
{
"name": {},
"value": {}
}
]
For AzureDatabricksDeltaLakeSink, use:
"type": "AzureDatabricksDeltaLakeSink",
"importSettings": {
"dateFormat": {},
"timestampFormat": {},
"type": "string",
"{customized property}": {}
},
"preCopyScript": {}
For AzureDataExplorerSink, use:
"type": "AzureDataExplorerSink",
"flushImmediately": {},
"ingestionMappingAsJson": {},
"ingestionMappingName": {}
For AzureDataLakeStoreSink, use:
"type": "AzureDataLakeStoreSink",
"copyBehavior": {},
"enableAdlsSingleFileParallel": {}
For AzureMySqlSink, use:
"type": "AzureMySqlSink",
"preCopyScript": {}
For AzurePostgreSqlSink, use:
"type": "AzurePostgreSqlSink",
"preCopyScript": {}
For AzureQueueSink, use:
"type": "AzureQueueSink"
For AzureSearchIndexSink, use:
"type": "AzureSearchIndexSink",
"writeBehavior": "string"
For AzureSqlSink, use:
"type": "AzureSqlSink",
"preCopyScript": {},
"sqlWriterStoredProcedureName": {},
"sqlWriterTableType": {},
"sqlWriterUseTableLock": {},
"storedProcedureParameters": {},
"storedProcedureTableTypeParameterName": {},
"tableOption": {},
"upsertSettings": {
"interimSchemaName": {},
"keys": {},
"useTempDB": {}
},
"writeBehavior": {}
For AzureTableSink, use:
"type": "AzureTableSink",
"azureTableDefaultPartitionKeyValue": {},
"azureTableInsertType": {},
"azureTablePartitionKeyName": {},
"azureTableRowKeyName": {}
For BinarySink, use:
"type": "BinarySink",
"storeSettings": {
"copyBehavior": {},
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"metadata": [
{
"name": {},
"value": {}
}
],
"type": "string"
// For remaining properties, see StoreWriteSettings objects
}
For BlobSink, use:
"type": "BlobSink",
"blobWriterAddHeader": {},
"blobWriterDateTimeFormat": {},
"blobWriterOverwriteFiles": {},
"copyBehavior": {},
"metadata": [
{
"name": {},
"value": {}
}
]
For CommonDataServiceForAppsSink, use:
"type": "CommonDataServiceForAppsSink",
"alternateKeyName": {},
"ignoreNullValues": {},
"writeBehavior": "Upsert"
For CosmosDbMongoDbApiSink, use:
"type": "CosmosDbMongoDbApiSink",
"writeBehavior": {}
For CosmosDbSqlApiSink, use:
"type": "CosmosDbSqlApiSink",
"writeBehavior": {}
For DelimitedTextSink, use:
"type": "DelimitedTextSink",
"formatSettings": {
"fileExtension": {},
"fileNamePrefix": {},
"maxRowsPerFile": {},
"quoteAllText": {},
"type": "string",
"{customized property}": {}
},
"storeSettings": {
"copyBehavior": {},
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"metadata": [
{
"name": {},
"value": {}
}
],
"type": "string"
// For remaining properties, see StoreWriteSettings objects
}
For DocumentDbCollectionSink, use:
"type": "DocumentDbCollectionSink",
"nestingSeparator": {},
"writeBehavior": {}
For DynamicsCrmSink, use:
"type": "DynamicsCrmSink",
"alternateKeyName": {},
"ignoreNullValues": {},
"writeBehavior": "Upsert"
For DynamicsSink, use:
"type": "DynamicsSink",
"alternateKeyName": {},
"ignoreNullValues": {},
"writeBehavior": "Upsert"
For FileSystemSink, use:
"type": "FileSystemSink",
"copyBehavior": {}
For InformixSink, use:
"type": "InformixSink",
"preCopyScript": {}
For JsonSink, use:
"type": "JsonSink",
"formatSettings": {
"filePattern": {},
"type": "string",
"{customized property}": {}
},
"storeSettings": {
"copyBehavior": {},
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"metadata": [
{
"name": {},
"value": {}
}
],
"type": "string"
// For remaining properties, see StoreWriteSettings objects
}
For LakeHouseTableSink, use:
"type": "LakeHouseTableSink",
"partitionNameList": {},
"partitionOption": {},
"tableActionOption": {}
For MicrosoftAccessSink, use:
"type": "MicrosoftAccessSink",
"preCopyScript": {}
For MongoDbAtlasSink, use:
"type": "MongoDbAtlasSink",
"writeBehavior": {}
For MongoDbV2Sink, use:
"type": "MongoDbV2Sink",
"writeBehavior": {}
For OdbcSink, use:
"type": "OdbcSink",
"preCopyScript": {}
For OracleSink, use:
"type": "OracleSink",
"preCopyScript": {}
For OrcSink, use:
"type": "OrcSink",
"formatSettings": {
"fileNamePrefix": {},
"maxRowsPerFile": {},
"type": "string",
"{customized property}": {}
},
"storeSettings": {
"copyBehavior": {},
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"metadata": [
{
"name": {},
"value": {}
}
],
"type": "string"
// For remaining properties, see StoreWriteSettings objects
}
For ParquetSink, use:
"type": "ParquetSink",
"formatSettings": {
"fileNamePrefix": {},
"maxRowsPerFile": {},
"type": "string",
"{customized property}": {}
},
"storeSettings": {
"copyBehavior": {},
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"metadata": [
{
"name": {},
"value": {}
}
],
"type": "string"
// For remaining properties, see StoreWriteSettings objects
}
For RestSink, use:
"type": "RestSink",
"additionalHeaders": {},
"httpCompressionType": {},
"httpRequestTimeout": {},
"requestInterval": {},
"requestMethod": {}
For SalesforceServiceCloudSink, use:
"type": "SalesforceServiceCloudSink",
"externalIdFieldName": {},
"ignoreNullValues": {},
"writeBehavior": "string"
For SalesforceServiceCloudV2Sink, use:
"type": "SalesforceServiceCloudV2Sink",
"externalIdFieldName": {},
"ignoreNullValues": {},
"writeBehavior": "string"
For SalesforceSink, use:
"type": "SalesforceSink",
"externalIdFieldName": {},
"ignoreNullValues": {},
"writeBehavior": "string"
For SalesforceV2Sink, use:
"type": "SalesforceV2Sink",
"externalIdFieldName": {},
"ignoreNullValues": {},
"writeBehavior": "string"
For SapCloudForCustomerSink, use:
"type": "SapCloudForCustomerSink",
"httpRequestTimeout": {},
"writeBehavior": "string"
For SnowflakeSink, use:
"type": "SnowflakeSink",
"importSettings": {
"additionalCopyOptions": {
"{customized property}": {}
},
"additionalFormatOptions": {
"{customized property}": {}
},
"type": "string",
"{customized property}": {}
},
"preCopyScript": {}
For SnowflakeV2Sink, use:
"type": "SnowflakeV2Sink",
"importSettings": {
"additionalCopyOptions": {
"{customized property}": {}
},
"additionalFormatOptions": {
"{customized property}": {}
},
"type": "string",
"{customized property}": {}
},
"preCopyScript": {}
For SqlDWSink, use:
"type": "SqlDWSink",
"allowCopyCommand": {},
"allowPolyBase": {},
"copyCommandSettings": {
"additionalOptions": {
"{customized property}": "string"
},
"defaultValues": [
{
"columnName": {},
"defaultValue": {}
}
]
},
"polyBaseSettings": {
"rejectSampleValue": {},
"rejectType": "string",
"rejectValue": {},
"useTypeDefault": {},
"{customized property}": {}
},
"preCopyScript": {},
"sqlWriterUseTableLock": {},
"tableOption": {},
"upsertSettings": {
"interimSchemaName": {},
"keys": {}
},
"writeBehavior": {}
For SqlMISink, use:
"type": "SqlMISink",
"preCopyScript": {},
"sqlWriterStoredProcedureName": {},
"sqlWriterTableType": {},
"sqlWriterUseTableLock": {},
"storedProcedureParameters": {},
"storedProcedureTableTypeParameterName": {},
"tableOption": {},
"upsertSettings": {
"interimSchemaName": {},
"keys": {},
"useTempDB": {}
},
"writeBehavior": {}
For SqlServerSink, use:
"type": "SqlServerSink",
"preCopyScript": {},
"sqlWriterStoredProcedureName": {},
"sqlWriterTableType": {},
"sqlWriterUseTableLock": {},
"storedProcedureParameters": {},
"storedProcedureTableTypeParameterName": {},
"tableOption": {},
"upsertSettings": {
"interimSchemaName": {},
"keys": {},
"useTempDB": {}
},
"writeBehavior": {}
For SqlSink, use:
"type": "SqlSink",
"preCopyScript": {},
"sqlWriterStoredProcedureName": {},
"sqlWriterTableType": {},
"sqlWriterUseTableLock": {},
"storedProcedureParameters": {},
"storedProcedureTableTypeParameterName": {},
"tableOption": {},
"upsertSettings": {
"interimSchemaName": {},
"keys": {},
"useTempDB": {}
},
"writeBehavior": {}
For WarehouseSink, use:
"type": "WarehouseSink",
"allowCopyCommand": {},
"copyCommandSettings": {
"additionalOptions": {
"{customized property}": "string"
},
"defaultValues": [
{
"columnName": {},
"defaultValue": {}
}
]
},
"preCopyScript": {},
"tableOption": {},
"writeBehavior": {}
StoreWriteSettings objects
Set the type property to specify the type of object.
For AzureBlobFSWriteSettings, use:
"type": "AzureBlobFSWriteSettings",
"blockSizeInMB": {}
For AzureBlobStorageWriteSettings, use:
"type": "AzureBlobStorageWriteSettings",
"blockSizeInMB": {}
For AzureDataLakeStoreWriteSettings, use:
"type": "AzureDataLakeStoreWriteSettings",
"expiryDateTime": {}
For AzureFileStorageWriteSettings, use:
"type": "AzureFileStorageWriteSettings"
For FileServerWriteSettings, use:
"type": "FileServerWriteSettings"
For LakeHouseWriteSettings, use:
"type": "LakeHouseWriteSettings"
For SftpWriteSettings, use:
"type": "SftpWriteSettings",
"operationTimeout": {},
"useTempFileRename": {}
CopySource objects
Set the type property to specify the type of object.
For AmazonMWSSource, use:
"type": "AmazonMWSSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For AmazonRdsForOracleSource, use:
"type": "AmazonRdsForOracleSource",
"additionalColumns": {},
"oracleReaderQuery": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionNames": {},
"partitionUpperBound": {}
},
"queryTimeout": {}
For AmazonRdsForSqlServerSource, use:
"type": "AmazonRdsForSqlServerSource",
"additionalColumns": {},
"isolationLevel": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"produceAdditionalTypes": {},
"queryTimeout": {},
"sqlReaderQuery": {},
"sqlReaderStoredProcedureName": {},
"storedProcedureParameters": {}
For AmazonRedshiftSource, use:
"type": "AmazonRedshiftSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {},
"redshiftUnloadSettings": {
"bucketName": {},
"s3LinkedServiceName": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
}
For AvroSource, use:
"type": "AvroSource",
"additionalColumns": {},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
For AzureBlobFSSource, use:
"type": "AzureBlobFSSource",
"recursive": {},
"skipHeaderLineCount": {},
"treatEmptyAsNull": {}
For AzureDatabricksDeltaLakeSource, use:
"type": "AzureDatabricksDeltaLakeSource",
"exportSettings": {
"dateFormat": {},
"timestampFormat": {},
"type": "string",
"{customized property}": {}
},
"query": {}
For AzureDataExplorerSource, use:
"type": "AzureDataExplorerSource",
"additionalColumns": {},
"noTruncation": {},
"query": {},
"queryTimeout": {}
For AzureDataLakeStoreSource, use:
"type": "AzureDataLakeStoreSource",
"recursive": {}
For AzureMariaDBSource, use:
"type": "AzureMariaDBSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For AzureMySqlSource, use:
"type": "AzureMySqlSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For AzurePostgreSqlSource, use:
"type": "AzurePostgreSqlSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For AzureSqlSource, use:
"type": "AzureSqlSource",
"additionalColumns": {},
"isolationLevel": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"produceAdditionalTypes": {},
"queryTimeout": {},
"sqlReaderQuery": {},
"sqlReaderStoredProcedureName": {},
"storedProcedureParameters": {}
For AzureTableSource, use:
"type": "AzureTableSource",
"additionalColumns": {},
"azureTableSourceIgnoreTableNotFound": {},
"azureTableSourceQuery": {},
"queryTimeout": {}
For BinarySource, use:
"type": "BinarySource",
"formatSettings": {
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
},
"type": "string",
"{customized property}": {}
},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
For BlobSource, use:
"type": "BlobSource",
"recursive": {},
"skipHeaderLineCount": {},
"treatEmptyAsNull": {}
For CassandraSource, use:
"type": "CassandraSource",
"additionalColumns": {},
"consistencyLevel": "string",
"query": {},
"queryTimeout": {}
For CommonDataServiceForAppsSource, use:
"type": "CommonDataServiceForAppsSource",
"additionalColumns": {},
"query": {}
For ConcurSource, use:
"type": "ConcurSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For CosmosDbMongoDbApiSource, use:
"type": "CosmosDbMongoDbApiSource",
"additionalColumns": {},
"batchSize": {},
"cursorMethods": {
"limit": {},
"project": {},
"skip": {},
"sort": {},
"{customized property}": {}
},
"filter": {},
"queryTimeout": {}
For CosmosDbSqlApiSource, use:
"type": "CosmosDbSqlApiSource",
"additionalColumns": {},
"detectDatetime": {},
"pageSize": {},
"preferredRegions": {},
"query": {}
For CouchbaseSource, use:
"type": "CouchbaseSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For Db2Source, use:
"type": "Db2Source",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For DelimitedTextSource, use:
"type": "DelimitedTextSource",
"additionalColumns": {},
"formatSettings": {
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
},
"skipLineCount": {},
"type": "string",
"{customized property}": {}
},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
For DocumentDbCollectionSource, use:
"type": "DocumentDbCollectionSource",
"additionalColumns": {},
"nestingSeparator": {},
"query": {},
"queryTimeout": {}
For DrillSource, use:
"type": "DrillSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For DynamicsAXSource, use:
"type": "DynamicsAXSource",
"additionalColumns": {},
"httpRequestTimeout": {},
"query": {},
"queryTimeout": {}
For DynamicsCrmSource, use:
"type": "DynamicsCrmSource",
"additionalColumns": {},
"query": {}
For DynamicsSource, use:
"type": "DynamicsSource",
"additionalColumns": {},
"query": {}
For EloquaSource, use:
"type": "EloquaSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For ExcelSource, use:
"type": "ExcelSource",
"additionalColumns": {},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
For FileSystemSource, use:
"type": "FileSystemSource",
"additionalColumns": {},
"recursive": {}
For GoogleAdWordsSource, use:
"type": "GoogleAdWordsSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For GoogleBigQuerySource, use:
"type": "GoogleBigQuerySource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For GoogleBigQueryV2Source, use:
"type": "GoogleBigQueryV2Source",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For GreenplumSource, use:
"type": "GreenplumSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For HBaseSource, use:
"type": "HBaseSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For HdfsSource, use:
"type": "HdfsSource",
"distcpSettings": {
"distcpOptions": {},
"resourceManagerEndpoint": {},
"tempScriptPath": {}
},
"recursive": {}
For HiveSource, use:
"type": "HiveSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For HttpSource, use:
"type": "HttpSource",
"httpRequestTimeout": {}
For HubspotSource, use:
"type": "HubspotSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For ImpalaSource, use:
"type": "ImpalaSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For InformixSource, use:
"type": "InformixSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For JiraSource, use:
"type": "JiraSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For JsonSource, use:
"type": "JsonSource",
"additionalColumns": {},
"formatSettings": {
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
},
"type": "string",
"{customized property}": {}
},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
For LakeHouseTableSource, use:
"type": "LakeHouseTableSource",
"additionalColumns": {},
"timestampAsOf": {},
"versionAsOf": {}
For MagentoSource, use:
"type": "MagentoSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For MariaDBSource, use:
"type": "MariaDBSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For MarketoSource, use:
"type": "MarketoSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For MicrosoftAccessSource, use:
"type": "MicrosoftAccessSource",
"additionalColumns": {},
"query": {}
For MongoDbAtlasSource, use:
"type": "MongoDbAtlasSource",
"additionalColumns": {},
"batchSize": {},
"cursorMethods": {
"limit": {},
"project": {},
"skip": {},
"sort": {},
"{customized property}": {}
},
"filter": {},
"queryTimeout": {}
For MongoDbSource, use:
"type": "MongoDbSource",
"additionalColumns": {},
"query": {}
For MongoDbV2Source, use:
"type": "MongoDbV2Source",
"additionalColumns": {},
"batchSize": {},
"cursorMethods": {
"limit": {},
"project": {},
"skip": {},
"sort": {},
"{customized property}": {}
},
"filter": {},
"queryTimeout": {}
For MySqlSource, use:
"type": "MySqlSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For NetezzaSource, use:
"type": "NetezzaSource",
"additionalColumns": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"query": {},
"queryTimeout": {}
For ODataSource, use:
"type": "ODataSource",
"additionalColumns": {},
"httpRequestTimeout": {},
"query": {}
For OdbcSource, use:
"type": "OdbcSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For Office365Source, use:
"type": "Office365Source",
"allowedGroups": {},
"dateFilterColumn": {},
"endTime": {},
"outputColumns": {},
"startTime": {},
"userScopeFilterUri": {}
For OracleServiceCloudSource, use:
"type": "OracleServiceCloudSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For OracleSource, use:
"type": "OracleSource",
"additionalColumns": {},
"oracleReaderQuery": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionNames": {},
"partitionUpperBound": {}
},
"queryTimeout": {}
For OrcSource, use:
"type": "OrcSource",
"additionalColumns": {},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
For ParquetSource, use:
"type": "ParquetSource",
"additionalColumns": {},
"formatSettings": {
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
},
"type": "string",
"{customized property}": {}
},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
For PaypalSource, use:
"type": "PaypalSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For PhoenixSource, use:
"type": "PhoenixSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For PostgreSqlSource, use:
"type": "PostgreSqlSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For PostgreSqlV2Source, use:
"type": "PostgreSqlV2Source",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For PrestoSource, use:
"type": "PrestoSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For QuickBooksSource, use:
"type": "QuickBooksSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For RelationalSource, use:
"type": "RelationalSource",
"additionalColumns": {},
"query": {}
For ResponsysSource, use:
"type": "ResponsysSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For RestSource, use:
"type": "RestSource",
"additionalColumns": {},
"additionalHeaders": {},
"httpRequestTimeout": {},
"paginationRules": {},
"requestBody": {},
"requestInterval": {},
"requestMethod": {}
For SalesforceMarketingCloudSource, use:
"type": "SalesforceMarketingCloudSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For SalesforceServiceCloudSource, use:
"type": "SalesforceServiceCloudSource",
"additionalColumns": {},
"query": {},
"readBehavior": {}
For SalesforceServiceCloudV2Source, use:
"type": "SalesforceServiceCloudV2Source",
"additionalColumns": {},
"includeDeletedObjects": {},
"SOQLQuery": {}
For SalesforceSource, use:
"type": "SalesforceSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {},
"readBehavior": {}
For SalesforceV2Source, use:
"type": "SalesforceV2Source",
"additionalColumns": {},
"includeDeletedObjects": {},
"queryTimeout": {},
"SOQLQuery": {}
For SapBwSource, use:
"type": "SapBwSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For SapCloudForCustomerSource, use:
"type": "SapCloudForCustomerSource",
"additionalColumns": {},
"httpRequestTimeout": {},
"query": {},
"queryTimeout": {}
For SapEccSource, use:
"type": "SapEccSource",
"additionalColumns": {},
"httpRequestTimeout": {},
"query": {},
"queryTimeout": {}
For SapHanaSource, use:
"type": "SapHanaSource",
"additionalColumns": {},
"packetSize": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {}
},
"query": {},
"queryTimeout": {}
For SapOdpSource, use:
"type": "SapOdpSource",
"additionalColumns": {},
"extractionMode": {},
"projection": {},
"queryTimeout": {},
"selection": {},
"subscriberProcess": {}
For SapOpenHubSource, use:
"type": "SapOpenHubSource",
"additionalColumns": {},
"baseRequestId": {},
"customRfcReadTableFunctionModule": {},
"excludeLastRequest": {},
"queryTimeout": {},
"sapDataColumnDelimiter": {}
For SapTableSource, use:
"type": "SapTableSource",
"additionalColumns": {},
"batchSize": {},
"customRfcReadTableFunctionModule": {},
"partitionOption": {},
"partitionSettings": {
"maxPartitionsNumber": {},
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"queryTimeout": {},
"rfcTableFields": {},
"rfcTableOptions": {},
"rowCount": {},
"rowSkips": {},
"sapDataColumnDelimiter": {}
For ServiceNowSource, use:
"type": "ServiceNowSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For ServiceNowV2Source, use:
"type": "ServiceNowV2Source",
"additionalColumns": {},
"expression": {
"operands": [
{
"operator": "string",
"type": "string",
"value": "string"
},
"queryTimeout": {}
For SharePointOnlineListSource, use:
"type": "SharePointOnlineListSource",
"httpRequestTimeout": {},
"query": {}
For ShopifySource, use:
"type": "ShopifySource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For SnowflakeSource, use:
"type": "SnowflakeSource",
"exportSettings": {
"additionalCopyOptions": {
"{customized property}": {}
},
"additionalFormatOptions": {
"{customized property}": {}
},
"type": "string",
"{customized property}": {}
},
"query": {}
For SnowflakeV2Source, use:
"type": "SnowflakeV2Source",
"exportSettings": {
"additionalCopyOptions": {
"{customized property}": {}
},
"additionalFormatOptions": {
"{customized property}": {}
},
"type": "string",
"{customized property}": {}
},
"query": {}
For SparkSource, use:
"type": "SparkSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For SqlDWSource, use:
"type": "SqlDWSource",
"additionalColumns": {},
"isolationLevel": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"queryTimeout": {},
"sqlReaderQuery": {},
"sqlReaderStoredProcedureName": {},
"storedProcedureParameters": {}
For SqlMISource, use:
"type": "SqlMISource",
"additionalColumns": {},
"isolationLevel": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"produceAdditionalTypes": {},
"queryTimeout": {},
"sqlReaderQuery": {},
"sqlReaderStoredProcedureName": {},
"storedProcedureParameters": {}
For SqlServerSource, use:
"type": "SqlServerSource",
"additionalColumns": {},
"isolationLevel": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"produceAdditionalTypes": {},
"queryTimeout": {},
"sqlReaderQuery": {},
"sqlReaderStoredProcedureName": {},
"storedProcedureParameters": {}
For SqlSource, use:
"type": "SqlSource",
"additionalColumns": {},
"isolationLevel": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"queryTimeout": {},
"sqlReaderQuery": {},
"sqlReaderStoredProcedureName": {},
"storedProcedureParameters": {}
For SquareSource, use:
"type": "SquareSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For SybaseSource, use:
"type": "SybaseSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For TeradataSource, use:
"type": "TeradataSource",
"additionalColumns": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"query": {},
"queryTimeout": {}
For VerticaSource, use:
"type": "VerticaSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For WarehouseSource, use:
"type": "WarehouseSource",
"additionalColumns": {},
"isolationLevel": {},
"partitionOption": {},
"partitionSettings": {
"partitionColumnName": {},
"partitionLowerBound": {},
"partitionUpperBound": {}
},
"queryTimeout": {},
"sqlReaderQuery": {},
"sqlReaderStoredProcedureName": {},
"storedProcedureParameters": {}
For WebSource, use:
"type": "WebSource",
"additionalColumns": {}
For XeroSource, use:
"type": "XeroSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
For XmlSource, use:
"type": "XmlSource",
"additionalColumns": {},
"formatSettings": {
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
},
"detectDataType": {},
"namespacePrefixes": {},
"namespaces": {},
"type": "string",
"validationMode": {},
"{customized property}": {}
},
"storeSettings": {
"disableMetricsCollection": {},
"maxConcurrentConnections": {},
"type": "string"
// For remaining properties, see StoreReadSettings objects
}
For ZohoSource, use:
"type": "ZohoSource",
"additionalColumns": {},
"query": {},
"queryTimeout": {}
StoreReadSettings objects
Set the type property to specify the type of object.
For AmazonS3CompatibleReadSettings, use:
"type": "AmazonS3CompatibleReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"prefix": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For AmazonS3ReadSettings, use:
"type": "AmazonS3ReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"prefix": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For AzureBlobFSReadSettings, use:
"type": "AzureBlobFSReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For AzureBlobStorageReadSettings, use:
"type": "AzureBlobStorageReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"prefix": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For AzureDataLakeStoreReadSettings, use:
"type": "AzureDataLakeStoreReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"listAfter": {},
"listBefore": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For AzureFileStorageReadSettings, use:
"type": "AzureFileStorageReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"prefix": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For FileServerReadSettings, use:
"type": "FileServerReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileFilter": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For FtpReadSettings, use:
"type": "FtpReadSettings",
"deleteFilesAfterCompletion": {},
"disableChunking": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"partitionRootPath": {},
"recursive": {},
"useBinaryTransfer": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For GoogleCloudStorageReadSettings, use:
"type": "GoogleCloudStorageReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"prefix": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For HdfsReadSettings, use:
"type": "HdfsReadSettings",
"deleteFilesAfterCompletion": {},
"distcpSettings": {
"distcpOptions": {},
"resourceManagerEndpoint": {},
"tempScriptPath": {}
},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For HttpReadSettings, use:
"type": "HttpReadSettings",
"additionalColumns": {},
"additionalHeaders": {},
"requestBody": {},
"requestMethod": {},
"requestTimeout": {}
For LakeHouseReadSettings, use:
"type": "LakeHouseReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For OracleCloudStorageReadSettings, use:
"type": "OracleCloudStorageReadSettings",
"deleteFilesAfterCompletion": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"prefix": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
For SftpReadSettings, use:
"type": "SftpReadSettings",
"deleteFilesAfterCompletion": {},
"disableChunking": {},
"enablePartitionDiscovery": {},
"fileListPath": {},
"modifiedDatetimeEnd": {},
"modifiedDatetimeStart": {},
"partitionRootPath": {},
"recursive": {},
"wildcardFileName": {},
"wildcardFolderPath": {}
CompressionReadSettings objects
Set the type property to specify the type of object.
For TarGZipReadSettings, use:
"type": "TarGZipReadSettings",
"preserveCompressionFileNameAsFolder": {}
For TarReadSettings, use:
"type": "TarReadSettings",
"preserveCompressionFileNameAsFolder": {}
For ZipDeflateReadSettings, use:
"type": "ZipDeflateReadSettings",
"preserveZipFileNameAsFolder": {}
SecretBase objects
Set the type property to specify the type of object.
For AzureKeyVaultSecret, use:
"type": "AzureKeyVaultSecret",
"secretName": {},
"secretVersion": {},
"store": {
"parameters": {
"{customized property}": {}
},
"referenceName": "string",
"type": "LinkedServiceReference"
}
For SecureString, use:
"type": "SecureString",
"value": "string"
FormatReadSettings objects
Set the type property to specify the type of object.
For BinaryReadSettings, use:
"type": "BinaryReadSettings",
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
}
For DelimitedTextReadSettings, use:
"type": "DelimitedTextReadSettings",
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
},
"skipLineCount": {}
For JsonReadSettings, use:
"type": "JsonReadSettings",
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
}
For ParquetReadSettings, use:
"type": "ParquetReadSettings",
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
}
For XmlReadSettings, use:
"type": "XmlReadSettings",
"compressionProperties": {
"type": "string"
// For remaining properties, see CompressionReadSettings objects
},
"detectDataType": {},
"namespacePrefixes": {},
"namespaces": {},
"validationMode": {}
Property values
factories/pipelines
Name | Description | Value |
---|---|---|
type | The resource type | 'Microsoft.DataFactory/factories/pipelines' |
apiVersion | The resource api version | '2018-06-01' |
name | The resource name See how to set names and types for child resources in JSON ARM templates. |
string (required) Character limit: 1-260 Valid characters: Can't use: <>*#.%&:\\+?/ or control charactersStart with alphanumeric. |
properties | Properties of the pipeline. | Pipeline (required) |
Pipeline
Name | Description | Value |
---|---|---|
activities | List of activities in pipeline. | Activity[] |
annotations | List of tags that can be used for describing the Pipeline. | any[] |
concurrency | The max number of concurrent runs for the pipeline. | int Constraints: Min value = 1 |
description | The description of the pipeline. | string |
folder | The folder that this Pipeline is in. If not specified, Pipeline will appear at the root level. | PipelineFolder |
parameters | List of parameters for pipeline. | ParameterDefinitionSpecification |
policy | Pipeline Policy. | PipelinePolicy |
runDimensions | Dimensions emitted by Pipeline. | PipelineRunDimensions |
variables | List of variables for pipeline. | VariableDefinitionSpecification |
Activity
Name | Description | Value |
---|---|---|
dependsOn | Activity depends on condition. | ActivityDependency[] |
description | Activity description. | string |
name | Activity name. | string (required) |
onInactiveMarkAs | Status result of the activity when the state is set to Inactive. This is an optional property and if not provided when the activity is inactive, the status will be Succeeded by default. | 'Failed' 'Skipped' 'Succeeded' |
state | Activity state. This is an optional property and if not provided, the state will be Active by default. | 'Active' 'Inactive' |
userProperties | Activity user properties. | UserProperty[] |
type | Set the object type | AppendVariable AzureDataExplorerCommand AzureFunctionActivity AzureMLBatchExecution AzureMLExecutePipeline AzureMLUpdateResource Copy Custom DatabricksNotebook DatabricksSparkJar DatabricksSparkPython DataLakeAnalyticsU-SQL Delete ExecuteDataFlow ExecutePipeline ExecuteSSISPackage ExecuteWranglingDataflow Fail Filter ForEach GetMetadata HDInsightHive HDInsightMapReduce HDInsightPig HDInsightSpark HDInsightStreaming IfCondition Lookup Script SetVariable SparkJob SqlServerStoredProcedure Switch SynapseNotebook Until Validation Wait WebActivity WebHook (required) |
ActivityDependency
Name | Description | Value |
---|---|---|
activity | Activity name. | string (required) |
dependencyConditions | Match-Condition for the dependency. | String array containing any of: 'Completed' 'Failed' 'Skipped' 'Succeeded' (required) |
{customized property} |
UserProperty
Name | Description | Value |
---|---|---|
name | User property name. | string (required) |
value | User property value. Type: string (or Expression with resultType string). |
AppendVariableActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'AppendVariable' (required) |
typeProperties | Append Variable activity properties. | AppendVariableActivityTypeProperties (required) |
AppendVariableActivityTypeProperties
Name | Description | Value |
---|---|---|
value | Value to be appended. Type: could be a static value matching type of the variable item or Expression with resultType matching type of the variable item | |
variableName | Name of the variable whose value needs to be appended to. | string |
AzureDataExplorerCommandActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'AzureDataExplorerCommand' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure Data Explorer command activity properties. | AzureDataExplorerCommandActivityTypeProperties (required) |
LinkedServiceReference
Name | Description | Value |
---|---|---|
parameters | Arguments for LinkedService. | ParameterValueSpecification |
referenceName | Reference LinkedService name. | string (required) |
type | Linked service reference type. | 'LinkedServiceReference' (required) |
ParameterValueSpecification
Name | Description | Value |
---|---|---|
{customized property} |
ActivityPolicy
Name | Description | Value |
---|---|---|
retry | Maximum ordinary retry attempts. Default is 0. Type: integer (or Expression with resultType integer), minimum: 0. | |
retryIntervalInSeconds | Interval between each retry attempt (in seconds). The default is 30 sec. | int Constraints: Min value = 30 Max value = 86400 |
secureInput | When set to true, Input from activity is considered as secure and will not be logged to monitoring. | bool |
secureOutput | When set to true, Output from activity is considered as secure and will not be logged to monitoring. | bool |
timeout | Specifies the timeout for the activity to run. The default timeout is 7 days. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
{customized property} |
AzureDataExplorerCommandActivityTypeProperties
Name | Description | Value |
---|---|---|
command | A control command, according to the Azure Data Explorer command syntax. Type: string (or Expression with resultType string). | |
commandTimeout | Control command timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9]))..) |
AzureFunctionActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'AzureFunctionActivity' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure Function activity properties. | AzureFunctionActivityTypeProperties (required) |
AzureFunctionActivityTypeProperties
Name | Description | Value |
---|---|---|
body | Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). | |
functionName | Name of the Function that the Azure Function Activity will call. Type: string (or Expression with resultType string) | |
headers | Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). | AzureFunctionActivityTypePropertiesHeaders |
method | Rest API method for target endpoint. | 'DELETE' 'GET' 'HEAD' 'OPTIONS' 'POST' 'PUT' 'TRACE' (required) |
AzureFunctionActivityTypePropertiesHeaders
Name | Description | Value |
---|---|---|
{customized property} | string |
AzureMLBatchExecutionActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'AzureMLBatchExecution' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure ML Batch Execution activity properties. | AzureMLBatchExecutionActivityTypeProperties (required) |
AzureMLBatchExecutionActivityTypeProperties
Name | Description | Value |
---|---|---|
globalParameters | Key,Value pairs to be passed to the Azure ML Batch Execution Service endpoint. Keys must match the names of web service parameters defined in the published Azure ML web service. Values will be passed in the GlobalParameters property of the Azure ML batch execution request. | AzureMLBatchExecutionActivityTypePropertiesGlobalPar... |
webServiceInputs | Key,Value pairs, mapping the names of Azure ML endpoint's Web Service Inputs to AzureMLWebServiceFile objects specifying the input Blob locations.. This information will be passed in the WebServiceInputs property of the Azure ML batch execution request. | AzureMLBatchExecutionActivityTypePropertiesWebServic... |
webServiceOutputs | Key,Value pairs, mapping the names of Azure ML endpoint's Web Service Outputs to AzureMLWebServiceFile objects specifying the output Blob locations. This information will be passed in the WebServiceOutputs property of the Azure ML batch execution request. | AzureMLBatchExecutionActivityTypePropertiesWebServic... |
AzureMLBatchExecutionActivityTypePropertiesGlobalPar...
Name | Description | Value |
---|---|---|
{customized property} |
AzureMLBatchExecutionActivityTypePropertiesWebServic...
Name | Description | Value |
---|---|---|
{customized property} | AzureMLWebServiceFile |
AzureMLWebServiceFile
Name | Description | Value |
---|---|---|
filePath | The relative file path, including container name, in the Azure Blob Storage specified by the LinkedService. Type: string (or Expression with resultType string). | |
linkedServiceName | Reference to an Azure Storage LinkedService, where Azure ML WebService Input/Output file located. | LinkedServiceReference (required) |
AzureMLBatchExecutionActivityTypePropertiesWebServic...
Name | Description | Value |
---|---|---|
{customized property} | AzureMLWebServiceFile |
AzureMLExecutePipelineActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'AzureMLExecutePipeline' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure ML Execute Pipeline activity properties. | AzureMLExecutePipelineActivityTypeProperties (required) |
AzureMLExecutePipelineActivityTypeProperties
Name | Description | Value |
---|---|---|
continueOnStepFailure | Whether to continue execution of other steps in the PipelineRun if a step fails. This information will be passed in the continueOnStepFailure property of the published pipeline execution request. Type: boolean (or Expression with resultType boolean). | |
dataPathAssignments | Dictionary used for changing data path assignments without retraining. Values will be passed in the dataPathAssignments property of the published pipeline execution request. Type: object (or Expression with resultType object). | |
experimentName | Run history experiment name of the pipeline run. This information will be passed in the ExperimentName property of the published pipeline execution request. Type: string (or Expression with resultType string). | |
mlParentRunId | The parent Azure ML Service pipeline run id. This information will be passed in the ParentRunId property of the published pipeline execution request. Type: string (or Expression with resultType string). | |
mlPipelineEndpointId | ID of the published Azure ML pipeline endpoint. Type: string (or Expression with resultType string). | |
mlPipelineId | ID of the published Azure ML pipeline. Type: string (or Expression with resultType string). | |
mlPipelineParameters | Key,Value pairs to be passed to the published Azure ML pipeline endpoint. Keys must match the names of pipeline parameters defined in the published pipeline. Values will be passed in the ParameterAssignments property of the published pipeline execution request. Type: object with key value pairs (or Expression with resultType object). | |
version | Version of the published Azure ML pipeline endpoint. Type: string (or Expression with resultType string). |
AzureMLUpdateResourceActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'AzureMLUpdateResource' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure ML Update Resource management activity properties. | AzureMLUpdateResourceActivityTypeProperties (required) |
AzureMLUpdateResourceActivityTypeProperties
Name | Description | Value |
---|---|---|
trainedModelFilePath | The relative file path in trainedModelLinkedService to represent the .ilearner file that will be uploaded by the update operation. Type: string (or Expression with resultType string). | |
trainedModelLinkedServiceName | Name of Azure Storage linked service holding the .ilearner file that will be uploaded by the update operation. | LinkedServiceReference (required) |
trainedModelName | Name of the Trained Model module in the Web Service experiment to be updated. Type: string (or Expression with resultType string). |
CopyActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Copy' (required) |
inputs | List of inputs for the activity. | DatasetReference[] |
linkedServiceName | Linked service reference. | LinkedServiceReference |
outputs | List of outputs for the activity. | DatasetReference[] |
policy | Activity policy. | ActivityPolicy |
typeProperties | Copy activity properties. | CopyActivityTypeProperties (required) |
DatasetReference
Name | Description | Value |
---|---|---|
parameters | Arguments for dataset. | ParameterValueSpecification |
referenceName | Reference dataset name. | string (required) |
type | Dataset reference type. | 'DatasetReference' (required) |
CopyActivityTypeProperties
Name | Description | Value |
---|---|---|
dataIntegrationUnits | Maximum number of data integration units that can be used to perform this data movement. Type: integer (or Expression with resultType integer), minimum: 0. | |
enableSkipIncompatibleRow | Whether to skip incompatible row. Default value is false. Type: boolean (or Expression with resultType boolean). | |
enableStaging | Specifies whether to copy data via an interim staging. Default value is false. Type: boolean (or Expression with resultType boolean). | |
logSettings | Log settings customer needs provide when enabling log. | LogSettings |
logStorageSettings | (Deprecated. Please use LogSettings) Log storage settings customer need to provide when enabling session log. | LogStorageSettings |
parallelCopies | Maximum number of concurrent sessions opened on the source or sink to avoid overloading the data store. Type: integer (or Expression with resultType integer), minimum: 0. | |
preserve | Preserve rules. | any[] |
preserveRules | Preserve Rules. | any[] |
redirectIncompatibleRowSettings | Redirect incompatible row settings when EnableSkipIncompatibleRow is true. | RedirectIncompatibleRowSettings |
sink | Copy activity sink. | CopySink (required) |
skipErrorFile | Specify the fault tolerance for data consistency. | SkipErrorFile |
source | Copy activity source. | CopySource (required) |
stagingSettings | Specifies interim staging settings when EnableStaging is true. | StagingSettings |
translator | Copy activity translator. If not specified, tabular translator is used. | |
validateDataConsistency | Whether to enable Data Consistency validation. Type: boolean (or Expression with resultType boolean). |
LogSettings
Name | Description | Value |
---|---|---|
copyActivityLogSettings | Specifies settings for copy activity log. | CopyActivityLogSettings |
enableCopyActivityLog | Specifies whether to enable copy activity log. Type: boolean (or Expression with resultType boolean). | |
logLocationSettings | Log location settings customer needs to provide when enabling log. | LogLocationSettings (required) |
CopyActivityLogSettings
Name | Description | Value |
---|---|---|
enableReliableLogging | Specifies whether to enable reliable logging. Type: boolean (or Expression with resultType boolean). | |
logLevel | Gets or sets the log level, support: Info, Warning. Type: string (or Expression with resultType string). |
LogLocationSettings
Name | Description | Value |
---|---|---|
linkedServiceName | Log storage linked service reference. | LinkedServiceReference (required) |
path | The path to storage for storing detailed logs of activity execution. Type: string (or Expression with resultType string). |
LogStorageSettings
Name | Description | Value |
---|---|---|
enableReliableLogging | Specifies whether to enable reliable logging. Type: boolean (or Expression with resultType boolean). | |
linkedServiceName | Log storage linked service reference. | LinkedServiceReference (required) |
logLevel | Gets or sets the log level, support: Info, Warning. Type: string (or Expression with resultType string). | |
path | The path to storage for storing detailed logs of activity execution. Type: string (or Expression with resultType string). | |
{customized property} |
RedirectIncompatibleRowSettings
Name | Description | Value |
---|---|---|
linkedServiceName | Name of the Azure Storage, Storage SAS, or Azure Data Lake Store linked service used for redirecting incompatible row. Must be specified if redirectIncompatibleRowSettings is specified. Type: string (or Expression with resultType string). | |
path | The path for storing the redirect incompatible row data. Type: string (or Expression with resultType string). | |
{customized property} |
CopySink
Name | Description | Value |
---|---|---|
disableMetricsCollection | If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). | |
maxConcurrentConnections | The maximum concurrent connection count for the sink data store. Type: integer (or Expression with resultType integer). | |
sinkRetryCount | Sink retry count. Type: integer (or Expression with resultType integer). | |
sinkRetryWait | Sink retry wait. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
writeBatchSize | Write batch size. Type: integer (or Expression with resultType integer), minimum: 0. | |
writeBatchTimeout | Write batch timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
type | Set the object type | AvroSink AzureBlobFSSink AzureDatabricksDeltaLakeSink AzureDataExplorerSink AzureDataLakeStoreSink AzureMySqlSink AzurePostgreSqlSink AzureQueueSink AzureSearchIndexSink AzureSqlSink AzureTableSink BinarySink BlobSink CommonDataServiceForAppsSink CosmosDbMongoDbApiSink CosmosDbSqlApiSink DelimitedTextSink DocumentDbCollectionSink DynamicsCrmSink DynamicsSink FileSystemSink InformixSink JsonSink LakeHouseTableSink MicrosoftAccessSink MongoDbAtlasSink MongoDbV2Sink OdbcSink OracleSink OrcSink ParquetSink RestSink SalesforceServiceCloudSink SalesforceServiceCloudV2Sink SalesforceSink SalesforceV2Sink SapCloudForCustomerSink SnowflakeSink SnowflakeV2Sink SqlDWSink SqlMISink SqlServerSink SqlSink WarehouseSink (required) |
AvroSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AvroSink' (required) |
formatSettings | Avro format settings. | AvroWriteSettings |
storeSettings | Avro store settings. | StoreWriteSettings |
AvroWriteSettings
Name | Description | Value |
---|---|---|
fileNamePrefix | Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). | |
maxRowsPerFile | Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). | |
recordName | Top level record name in write result, which is required in AVRO spec. | string |
recordNamespace | Record namespace in the write result. | string |
type | The write setting type. | string (required) |
{customized property} |
StoreWriteSettings
Name | Description | Value |
---|---|---|
copyBehavior | The type of copy behavior for copy sink. | |
disableMetricsCollection | If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). | |
maxConcurrentConnections | The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). | |
metadata | Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). | MetadataItem[] |
type | Set the object type | AzureBlobFSWriteSettings AzureBlobStorageWriteSettings AzureDataLakeStoreWriteSettings AzureFileStorageWriteSettings FileServerWriteSettings LakeHouseWriteSettings SftpWriteSettings (required) |
MetadataItem
Name | Description | Value |
---|---|---|
name | Metadata item key name. Type: string (or Expression with resultType string). | |
value | Metadata item value. Type: string (or Expression with resultType string). |
AzureBlobFSWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | 'AzureBlobFSWriteSettings' (required) |
blockSizeInMB | Indicates the block size(MB) when writing data to blob. Type: integer (or Expression with resultType integer). |
AzureBlobStorageWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | 'AzureBlobStorageWriteSettings' (required) |
blockSizeInMB | Indicates the block size(MB) when writing data to blob. Type: integer (or Expression with resultType integer). |
AzureDataLakeStoreWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | 'AzureDataLakeStoreWriteSettings' (required) |
expiryDateTime | Specifies the expiry time of the written files. The time is applied to the UTC time zone in the format of "2018-12-01T05:00:00Z". Default value is NULL. Type: string (or Expression with resultType string). |
AzureFileStorageWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | 'AzureFileStorageWriteSettings' (required) |
FileServerWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | 'FileServerWriteSettings' (required) |
LakeHouseWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | 'LakeHouseWriteSettings' (required) |
SftpWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | 'SftpWriteSettings' (required) |
operationTimeout | Specifies the timeout for writing each chunk to SFTP server. Default value: 01:00:00 (one hour). Type: string (or Expression with resultType string). | |
useTempFileRename | Upload to temporary file(s) and rename. Disable this option if your SFTP server doesn't support rename operation. Type: boolean (or Expression with resultType boolean). |
AzureBlobFSSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureBlobFSSink' (required) |
copyBehavior | The type of copy behavior for copy sink. Type: string (or Expression with resultType string). | |
metadata | Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). | MetadataItem[] |
AzureDatabricksDeltaLakeSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureDatabricksDeltaLakeSink' (required) |
importSettings | Azure Databricks Delta Lake import settings. | AzureDatabricksDeltaLakeImportCommand |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). |
AzureDatabricksDeltaLakeImportCommand
Name | Description | Value |
---|---|---|
dateFormat | Specify the date format for csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). | |
timestampFormat | Specify the timestamp format for csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). | |
type | The import setting type. | string (required) |
{customized property} |
AzureDataExplorerSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureDataExplorerSink' (required) |
flushImmediately | If set to true, any aggregation will be skipped. Default is false. Type: boolean. | |
ingestionMappingAsJson | An explicit column mapping description provided in a json format. Type: string. | |
ingestionMappingName | A name of a pre-created csv mapping that was defined on the target Kusto table. Type: string. |
AzureDataLakeStoreSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureDataLakeStoreSink' (required) |
copyBehavior | The type of copy behavior for copy sink. Type: string (or Expression with resultType string). | |
enableAdlsSingleFileParallel | Single File Parallel. |
AzureMySqlSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureMySqlSink' (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). |
AzurePostgreSqlSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzurePostgreSqlSink' (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). |
AzureQueueSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureQueueSink' (required) |
AzureSearchIndexSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureSearchIndexSink' (required) |
writeBehavior | Specify the write behavior when upserting documents into Azure Search Index. | 'Merge' 'Upload' |
AzureSqlSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureSqlSink' (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | |
sqlWriterStoredProcedureName | SQL writer stored procedure name. Type: string (or Expression with resultType string). | |
sqlWriterTableType | SQL writer table type. Type: string (or Expression with resultType string). | |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | |
storedProcedureParameters | SQL stored procedure parameters. | |
storedProcedureTableTypeParameterName | The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). | |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | |
upsertSettings | SQL upsert settings. | SqlUpsertSettings |
writeBehavior | Write behavior when copying data into Azure SQL. Type: SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum) |
SqlUpsertSettings
Name | Description | Value |
---|---|---|
interimSchemaName | Schema name for interim table. Type: string (or Expression with resultType string). | |
keys | Key column names for unique row identification. Type: array of strings (or Expression with resultType array of strings). | |
useTempDB | Specifies whether to use temp db for upsert interim table. Type: boolean (or Expression with resultType boolean). |
AzureTableSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'AzureTableSink' (required) |
azureTableDefaultPartitionKeyValue | Azure Table default partition key value. Type: string (or Expression with resultType string). | |
azureTableInsertType | Azure Table insert type. Type: string (or Expression with resultType string). | |
azureTablePartitionKeyName | Azure Table partition key name. Type: string (or Expression with resultType string). | |
azureTableRowKeyName | Azure Table row key name. Type: string (or Expression with resultType string). |
BinarySink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'BinarySink' (required) |
storeSettings | Binary store settings. | StoreWriteSettings |
BlobSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'BlobSink' (required) |
blobWriterAddHeader | Blob writer add header. Type: boolean (or Expression with resultType boolean). | |
blobWriterDateTimeFormat | Blob writer date time format. Type: string (or Expression with resultType string). | |
blobWriterOverwriteFiles | Blob writer overwrite files. Type: boolean (or Expression with resultType boolean). | |
copyBehavior | The type of copy behavior for copy sink. | |
metadata | Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). | MetadataItem[] |
CommonDataServiceForAppsSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'CommonDataServiceForAppsSink' (required) |
alternateKeyName | The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether to ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. | 'Upsert' (required) |
CosmosDbMongoDbApiSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'CosmosDbMongoDbApiSink' (required) |
writeBehavior | Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). |
CosmosDbSqlApiSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'CosmosDbSqlApiSink' (required) |
writeBehavior | Describes how to write data to Azure Cosmos DB. Type: string (or Expression with resultType string). Allowed values: insert and upsert. |
DelimitedTextSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'DelimitedTextSink' (required) |
formatSettings | DelimitedText format settings. | DelimitedTextWriteSettings |
storeSettings | DelimitedText store settings. | StoreWriteSettings |
DelimitedTextWriteSettings
Name | Description | Value |
---|---|---|
fileExtension | The file extension used to create the files. Type: string (or Expression with resultType string). | |
fileNamePrefix | Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). | |
maxRowsPerFile | Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). | |
quoteAllText | Indicates whether string values should always be enclosed with quotes. Type: boolean (or Expression with resultType boolean). | |
type | The write setting type. | string (required) |
{customized property} |
DocumentDbCollectionSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'DocumentDbCollectionSink' (required) |
nestingSeparator | Nested properties separator. Default is . (dot). Type: string (or Expression with resultType string). | |
writeBehavior | Describes how to write data to Azure Cosmos DB. Type: string (or Expression with resultType string). Allowed values: insert and upsert. |
DynamicsCrmSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'DynamicsCrmSink' (required) |
alternateKeyName | The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether to ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. | 'Upsert' (required) |
DynamicsSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'DynamicsSink' (required) |
alternateKeyName | The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. | 'Upsert' (required) |
FileSystemSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'FileSystemSink' (required) |
copyBehavior | The type of copy behavior for copy sink. |
InformixSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'InformixSink' (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). |
JsonSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'JsonSink' (required) |
formatSettings | Json format settings. | JsonWriteSettings |
storeSettings | Json store settings. | StoreWriteSettings |
JsonWriteSettings
Name | Description | Value |
---|---|---|
filePattern | File pattern of JSON. This setting controls the way a collection of JSON objects will be treated. The default value is 'setOfObjects'. It is case-sensitive. | |
type | The write setting type. | string (required) |
{customized property} |
LakeHouseTableSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'LakeHouseTableSink' (required) |
partitionNameList | Specify the partition column names from sink columns. Type: array of objects (or Expression with resultType array of objects). | |
partitionOption | Create partitions in folder structure based on one or multiple columns. Each distinct column value (pair) will be a new partition. Possible values include: "None", "PartitionByKey". | |
tableActionOption | The type of table action for LakeHouse Table sink. Possible values include: "None", "Append", "Overwrite". |
MicrosoftAccessSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'MicrosoftAccessSink' (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). |
MongoDbAtlasSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'MongoDbAtlasSink' (required) |
writeBehavior | Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). |
MongoDbV2Sink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'MongoDbV2Sink' (required) |
writeBehavior | Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). |
OdbcSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'OdbcSink' (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). |
OracleSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'OracleSink' (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). |
OrcSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'OrcSink' (required) |
formatSettings | ORC format settings. | OrcWriteSettings |
storeSettings | ORC store settings. | StoreWriteSettings |
OrcWriteSettings
Name | Description | Value |
---|---|---|
fileNamePrefix | Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). | |
maxRowsPerFile | Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). | |
type | The write setting type. | string (required) |
{customized property} |
ParquetSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'ParquetSink' (required) |
formatSettings | Parquet format settings. | ParquetWriteSettings |
storeSettings | Parquet store settings. | StoreWriteSettings |
ParquetWriteSettings
Name | Description | Value |
---|---|---|
fileNamePrefix | Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). | |
maxRowsPerFile | Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). | |
type | The write setting type. | string (required) |
{customized property} |
RestSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'RestSink' (required) |
additionalHeaders | The additional HTTP headers in the request to the RESTful API. Type: key value pairs (value should be string type). | |
httpCompressionType | Http Compression Type to Send data in compressed format with Optimal Compression Level, Default is None. And The Only Supported option is Gzip. Type: string (or Expression with resultType string). | |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
requestInterval | The time to await before sending next request, in milliseconds | |
requestMethod | The HTTP method used to call the RESTful API. The default is POST. Type: string (or Expression with resultType string). |
SalesforceServiceCloudSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SalesforceServiceCloudSink' (required) |
externalIdFieldName | The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. Default is Insert. | 'Insert' 'Upsert' |
SalesforceServiceCloudV2Sink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SalesforceServiceCloudV2Sink' (required) |
externalIdFieldName | The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. Default is Insert. | 'Insert' 'Upsert' |
SalesforceSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SalesforceSink' (required) |
externalIdFieldName | The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. Default is Insert. | 'Insert' 'Upsert' |
SalesforceV2Sink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SalesforceV2Sink' (required) |
externalIdFieldName | The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. Default is Insert. | 'Insert' 'Upsert' |
SapCloudForCustomerSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SapCloudForCustomerSink' (required) |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
writeBehavior | The write behavior for the operation. Default is 'Insert'. | 'Insert' 'Update' |
SnowflakeSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SnowflakeSink' (required) |
importSettings | Snowflake import settings. | SnowflakeImportCopyCommand |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). |
SnowflakeImportCopyCommand
Name | Description | Value |
---|---|---|
additionalCopyOptions | Additional copy options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": "'HH24:MI:SS.FF'" } | SnowflakeImportCopyCommandAdditionalCopyOptions |
additionalFormatOptions | Additional format options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalFormatOptions": { "FORCE": "TRUE", "LOAD_UNCERTAIN_FILES": "'FALSE'" } | SnowflakeImportCopyCommandAdditionalFormatOptions |
type | The import setting type. | string (required) |
{customized property} |
SnowflakeImportCopyCommandAdditionalCopyOptions
Name | Description | Value |
---|---|---|
{customized property} |
SnowflakeImportCopyCommandAdditionalFormatOptions
Name | Description | Value |
---|---|---|
{customized property} |
SnowflakeV2Sink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SnowflakeV2Sink' (required) |
importSettings | Snowflake import settings. | SnowflakeImportCopyCommand |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). |
SqlDWSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SqlDWSink' (required) |
allowCopyCommand | Indicates to use Copy Command to copy data into SQL Data Warehouse. Type: boolean (or Expression with resultType boolean). | |
allowPolyBase | Indicates to use PolyBase to copy data into SQL Data Warehouse when applicable. Type: boolean (or Expression with resultType boolean). | |
copyCommandSettings | Specifies Copy Command related settings when allowCopyCommand is true. | DWCopyCommandSettings |
polyBaseSettings | Specifies PolyBase-related settings when allowPolyBase is true. | PolybaseSettings |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | |
upsertSettings | SQL DW upsert settings. | SqlDWUpsertSettings |
writeBehavior | Write behavior when copying data into azure SQL DW. Type: SqlDWWriteBehaviorEnum (or Expression with resultType SqlDWWriteBehaviorEnum) |
DWCopyCommandSettings
Name | Description | Value |
---|---|---|
additionalOptions | Additional options directly passed to SQL DW in Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalOptions": { "MAXERRORS": "1000", "DATEFORMAT": "'ymd'" } | DWCopyCommandSettingsAdditionalOptions |
defaultValues | Specifies the default values for each target column in SQL DW. The default values in the property overwrite the DEFAULT constraint set in the DB, and identity column cannot have a default value. Type: array of objects (or Expression with resultType array of objects). | DWCopyCommandDefaultValue[] |
DWCopyCommandSettingsAdditionalOptions
Name | Description | Value |
---|---|---|
{customized property} | string |
DWCopyCommandDefaultValue
Name | Description | Value |
---|---|---|
columnName | Column name. Type: object (or Expression with resultType string). | |
defaultValue | The default value of the column. Type: object (or Expression with resultType string). |
PolybaseSettings
Name | Description | Value |
---|---|---|
rejectSampleValue | Determines the number of rows to attempt to retrieve before the PolyBase recalculates the percentage of rejected rows. Type: integer (or Expression with resultType integer), minimum: 0. | |
rejectType | Reject type. | 'percentage' 'value' |
rejectValue | Specifies the value or the percentage of rows that can be rejected before the query fails. Type: number (or Expression with resultType number), minimum: 0. | |
useTypeDefault | Specifies how to handle missing values in delimited text files when PolyBase retrieves data from the text file. Type: boolean (or Expression with resultType boolean). | |
{customized property} |
SqlDWUpsertSettings
Name | Description | Value |
---|---|---|
interimSchemaName | Schema name for interim table. Type: string (or Expression with resultType string). | |
keys | Key column names for unique row identification. Type: array of strings (or Expression with resultType array of strings). |
SqlMISink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SqlMISink' (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | |
sqlWriterStoredProcedureName | SQL writer stored procedure name. Type: string (or Expression with resultType string). | |
sqlWriterTableType | SQL writer table type. Type: string (or Expression with resultType string). | |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | |
storedProcedureParameters | SQL stored procedure parameters. | |
storedProcedureTableTypeParameterName | The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). | |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | |
upsertSettings | SQL upsert settings. | SqlUpsertSettings |
writeBehavior | White behavior when copying data into azure SQL MI. Type: string (or Expression with resultType string) |
SqlServerSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SqlServerSink' (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | |
sqlWriterStoredProcedureName | SQL writer stored procedure name. Type: string (or Expression with resultType string). | |
sqlWriterTableType | SQL writer table type. Type: string (or Expression with resultType string). | |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | |
storedProcedureParameters | SQL stored procedure parameters. | |
storedProcedureTableTypeParameterName | The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). | |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | |
upsertSettings | SQL upsert settings. | SqlUpsertSettings |
writeBehavior | Write behavior when copying data into sql server. Type: string (or Expression with resultType string). |
SqlSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'SqlSink' (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | |
sqlWriterStoredProcedureName | SQL writer stored procedure name. Type: string (or Expression with resultType string). | |
sqlWriterTableType | SQL writer table type. Type: string (or Expression with resultType string). | |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | |
storedProcedureParameters | SQL stored procedure parameters. | |
storedProcedureTableTypeParameterName | The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). | |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | |
upsertSettings | SQL upsert settings. | SqlUpsertSettings |
writeBehavior | Write behavior when copying data into sql. Type: string (or Expression with resultType string). |
WarehouseSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | 'WarehouseSink' (required) |
allowCopyCommand | Indicates to use Copy Command to copy data into SQL Data Warehouse. Type: boolean (or Expression with resultType boolean). | |
copyCommandSettings | Specifies Copy Command related settings when allowCopyCommand is true. | DWCopyCommandSettings |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | |
writeBehavior | Write behavior when copying data into azure Microsoft Fabric Data Warehouse. Type: DWWriteBehaviorEnum (or Expression with resultType DWWriteBehaviorEnum) |
SkipErrorFile
Name | Description | Value |
---|---|---|
dataInconsistency | Skip if source/sink file changed by other concurrent write. Default is false. Type: boolean (or Expression with resultType boolean). | |
fileMissing | Skip if file is deleted by other client during copy. Default is true. Type: boolean (or Expression with resultType boolean). |
CopySource
AmazonMWSSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AmazonMWSSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
AmazonRdsForOracleSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AmazonRdsForOracleSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
oracleReaderQuery | AmazonRdsForOracle reader query. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for AmazonRdsForOracle read in parallel. Type: string (or Expression with resultType string). | |
partitionSettings | The settings that will be leveraged for AmazonRdsForOracle source partitioning. | AmazonRdsForOraclePartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
AmazonRdsForOraclePartitionSettings
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionNames | Names of the physical partitions of AmazonRdsForOracle table. | |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). |
AmazonRdsForSqlServerSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AmazonRdsForSqlServerSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
produceAdditionalTypes | Which additional types to produce. | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". |
SqlPartitionSettings
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column in integer or datetime type that will be used for proceeding partitioning. If not specified, the primary key of the table is auto-detected and used as the partition column. Type: string (or Expression with resultType string). | |
partitionLowerBound | The minimum value of the partition column for partition range splitting. This value is used to decide the partition stride, not for filtering the rows in table. All rows in the table or query result will be partitioned and copied. Type: string (or Expression with resultType string). | |
partitionUpperBound | The maximum value of the partition column for partition range splitting. This value is used to decide the partition stride, not for filtering the rows in table. All rows in the table or query result will be partitioned and copied. Type: string (or Expression with resultType string). |
AmazonRedshiftSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AmazonRedshiftSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
redshiftUnloadSettings | The Amazon S3 settings needed for the interim Amazon S3 when copying from Amazon Redshift with unload. With this, data from Amazon Redshift source will be unloaded into S3 first and then copied into the targeted sink from the interim S3. | RedshiftUnloadSettings |
RedshiftUnloadSettings
Name | Description | Value |
---|---|---|
bucketName | The bucket of the interim Amazon S3 which will be used to store the unloaded data from Amazon Redshift source. The bucket must be in the same region as the Amazon Redshift source. Type: string (or Expression with resultType string). | |
s3LinkedServiceName | The name of the Amazon S3 linked service which will be used for the unload operation when copying from the Amazon Redshift source. | LinkedServiceReference (required) |
AvroSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AvroSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
storeSettings | Avro store settings. | StoreReadSettings |
StoreReadSettings
Name | Description | Value |
---|---|---|
disableMetricsCollection | If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). | |
maxConcurrentConnections | The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). | |
type | Set the object type | AmazonS3CompatibleReadSettings AmazonS3ReadSettings AzureBlobFSReadSettings AzureBlobStorageReadSettings AzureDataLakeStoreReadSettings AzureFileStorageReadSettings FileServerReadSettings FtpReadSettings GoogleCloudStorageReadSettings HdfsReadSettings HttpReadSettings LakeHouseReadSettings OracleCloudStorageReadSettings SftpReadSettings (required) |
AmazonS3CompatibleReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'AmazonS3CompatibleReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
prefix | The prefix filter for the S3 Compatible object name. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Amazon S3 Compatible wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Amazon S3 Compatible wildcardFolderPath. Type: string (or Expression with resultType string). |
AmazonS3ReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'AmazonS3ReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
prefix | The prefix filter for the S3 object name. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | AmazonS3 wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | AmazonS3 wildcardFolderPath. Type: string (or Expression with resultType string). |
AzureBlobFSReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'AzureBlobFSReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Azure blobFS wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Azure blobFS wildcardFolderPath. Type: string (or Expression with resultType string). |
AzureBlobStorageReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'AzureBlobStorageReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
prefix | The prefix filter for the Azure Blob name. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Azure blob wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Azure blob wildcardFolderPath. Type: string (or Expression with resultType string). |
AzureDataLakeStoreReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'AzureDataLakeStoreReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
listAfter | Lists files after the value (exclusive) based on file/folder names’ lexicographical order. Applies under the folderPath in data set, and filter files/sub-folders under the folderPath. Type: string (or Expression with resultType string). | |
listBefore | Lists files before the value (inclusive) based on file/folder names’ lexicographical order. Applies under the folderPath in data set, and filter files/sub-folders under the folderPath. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | ADLS wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | ADLS wildcardFolderPath. Type: string (or Expression with resultType string). |
AzureFileStorageReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'AzureFileStorageReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
prefix | The prefix filter for the Azure File name starting from root path. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Azure File Storage wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Azure File Storage wildcardFolderPath. Type: string (or Expression with resultType string). |
FileServerReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'FileServerReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileFilter | Specify a filter to be used to select a subset of files in the folderPath rather than all files. Type: string (or Expression with resultType string). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | FileServer wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | FileServer wildcardFolderPath. Type: string (or Expression with resultType string). |
FtpReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'FtpReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
disableChunking | If true, disable parallel reading within each file. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
useBinaryTransfer | Specify whether to use binary transfer mode for FTP stores. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Ftp wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Ftp wildcardFolderPath. Type: string (or Expression with resultType string). |
GoogleCloudStorageReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'GoogleCloudStorageReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
prefix | The prefix filter for the Google Cloud Storage object name. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Google Cloud Storage wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Google Cloud Storage wildcardFolderPath. Type: string (or Expression with resultType string). |
HdfsReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'HdfsReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
distcpSettings | Specifies Distcp-related settings. | DistcpSettings |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | HDFS wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | HDFS wildcardFolderPath. Type: string (or Expression with resultType string). |
DistcpSettings
Name | Description | Value |
---|---|---|
distcpOptions | Specifies the Distcp options. Type: string (or Expression with resultType string). | |
resourceManagerEndpoint | Specifies the Yarn ResourceManager endpoint. Type: string (or Expression with resultType string). | |
tempScriptPath | Specifies an existing folder path which will be used to store temp Distcp command script. The script file is generated by ADF and will be removed after Copy job finished. Type: string (or Expression with resultType string). |
HttpReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'HttpReadSettings' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
additionalHeaders | The additional HTTP headers in the request to the RESTful API. Type: string (or Expression with resultType string). | |
requestBody | The HTTP request body to the RESTful API if requestMethod is POST. Type: string (or Expression with resultType string). | |
requestMethod | The HTTP method used to call the RESTful API. The default is GET. Type: string (or Expression with resultType string). | |
requestTimeout | Specifies the timeout for a HTTP client to get HTTP response from HTTP server. Type: string (or Expression with resultType string). |
LakeHouseReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'LakeHouseReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Microsoft Fabric LakeHouse Files wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Microsoft Fabric LakeHouse Files wildcardFolderPath. Type: string (or Expression with resultType string). |
OracleCloudStorageReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'OracleCloudStorageReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
prefix | The prefix filter for the Oracle Cloud Storage object name. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Oracle Cloud Storage wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Oracle Cloud Storage wildcardFolderPath. Type: string (or Expression with resultType string). |
SftpReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | 'SftpReadSettings' (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
disableChunking | If true, disable parallel reading within each file. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Sftp wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Sftp wildcardFolderPath. Type: string (or Expression with resultType string). |
AzureBlobFSSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureBlobFSSource' (required) |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
skipHeaderLineCount | Number of header lines to skip from each blob. Type: integer (or Expression with resultType integer). | |
treatEmptyAsNull | Treat empty as null. Type: boolean (or Expression with resultType boolean). |
AzureDatabricksDeltaLakeSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureDatabricksDeltaLakeSource' (required) |
exportSettings | Azure Databricks Delta Lake export settings. | AzureDatabricksDeltaLakeExportCommand |
query | Azure Databricks Delta Lake Sql query. Type: string (or Expression with resultType string). |
AzureDatabricksDeltaLakeExportCommand
Name | Description | Value |
---|---|---|
dateFormat | Specify the date format for the csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). | |
timestampFormat | Specify the timestamp format for the csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). | |
type | The export setting type. | string (required) |
{customized property} |
AzureDataExplorerSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureDataExplorerSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
noTruncation | The name of the Boolean option that controls whether truncation is applied to result-sets that go beyond a certain row-count limit. | |
query | Database query. Should be a Kusto Query Language (KQL) query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).. |
AzureDataLakeStoreSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureDataLakeStoreSource' (required) |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). |
AzureMariaDBSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureMariaDBSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
AzureMySqlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureMySqlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
AzurePostgreSqlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzurePostgreSqlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
AzureSqlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureSqlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
produceAdditionalTypes | Which additional types to produce. | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". |
AzureTableSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'AzureTableSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
azureTableSourceIgnoreTableNotFound | Azure Table source ignore table not found. Type: boolean (or Expression with resultType boolean). | |
azureTableSourceQuery | Azure Table source query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
BinarySource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'BinarySource' (required) |
formatSettings | Binary format settings. | BinaryReadSettings |
storeSettings | Binary store settings. | StoreReadSettings |
BinaryReadSettings
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
type | The read setting type. | string (required) |
{customized property} |
CompressionReadSettings
Name | Description | Value |
---|---|---|
type | Set the object type | TarGZipReadSettings TarReadSettings ZipDeflateReadSettings (required) |
TarGZipReadSettings
Name | Description | Value |
---|---|---|
type | The Compression setting type. | 'TarGZipReadSettings' (required) |
preserveCompressionFileNameAsFolder | Preserve the compression file name as folder path. Type: boolean (or Expression with resultType boolean). |
TarReadSettings
Name | Description | Value |
---|---|---|
type | The Compression setting type. | 'TarReadSettings' (required) |
preserveCompressionFileNameAsFolder | Preserve the compression file name as folder path. Type: boolean (or Expression with resultType boolean). |
ZipDeflateReadSettings
Name | Description | Value |
---|---|---|
type | The Compression setting type. | 'ZipDeflateReadSettings' (required) |
preserveZipFileNameAsFolder | Preserve the zip file name as folder path. Type: boolean (or Expression with resultType boolean). |
BlobSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'BlobSource' (required) |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
skipHeaderLineCount | Number of header lines to skip from each blob. Type: integer (or Expression with resultType integer). | |
treatEmptyAsNull | Treat empty as null. Type: boolean (or Expression with resultType boolean). |
CassandraSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'CassandraSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
consistencyLevel | The consistency level specifies how many Cassandra servers must respond to a read request before returning data to the client application. Cassandra checks the specified number of Cassandra servers for data to satisfy the read request. Must be one of cassandraSourceReadConsistencyLevels. The default value is 'ONE'. It is case-insensitive. | 'ALL' 'EACH_QUORUM' 'LOCAL_ONE' 'LOCAL_QUORUM' 'LOCAL_SERIAL' 'ONE' 'QUORUM' 'SERIAL' 'THREE' 'TWO' |
query | Database query. Should be a SQL-92 query expression or Cassandra Query Language (CQL) command. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
CommonDataServiceForAppsSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'CommonDataServiceForAppsSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | FetchXML is a proprietary query language that is used in Microsoft Common Data Service for Apps (online & on-premises). Type: string (or Expression with resultType string). |
ConcurSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ConcurSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
CosmosDbMongoDbApiSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'CosmosDbMongoDbApiSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
batchSize | Specifies the number of documents to return in each batch of the response from MongoDB instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer). | |
cursorMethods | Cursor methods for Mongodb query. | MongoDbCursorMethodsProperties |
filter | Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
MongoDbCursorMethodsProperties
Name | Description | Value |
---|---|---|
limit | Specifies the maximum number of documents the server returns. limit() is analogous to the LIMIT statement in a SQL database. Type: integer (or Expression with resultType integer). | |
project | Specifies the fields to return in the documents that match the query filter. To return all fields in the matching documents, omit this parameter. Type: string (or Expression with resultType string). | |
skip | Specifies the how many documents skipped and where MongoDB begins returning results. This approach may be useful in implementing paginated results. Type: integer (or Expression with resultType integer). | |
sort | Specifies the order in which the query returns matching documents. Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). | |
{customized property} |
CosmosDbSqlApiSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'CosmosDbSqlApiSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
detectDatetime | Whether detect primitive values as datetime values. Type: boolean (or Expression with resultType boolean). | |
pageSize | Page size of the result. Type: integer (or Expression with resultType integer). | |
preferredRegions | Preferred regions. Type: array of strings (or Expression with resultType array of strings). | |
query | SQL API query. Type: string (or Expression with resultType string). |
CouchbaseSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'CouchbaseSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Db2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | 'Db2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
DelimitedTextSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'DelimitedTextSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
formatSettings | DelimitedText format settings. | DelimitedTextReadSettings |
storeSettings | DelimitedText store settings. | StoreReadSettings |
DelimitedTextReadSettings
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
skipLineCount | Indicates the number of non-empty rows to skip when reading data from input files. Type: integer (or Expression with resultType integer). | |
type | The read setting type. | string (required) |
{customized property} |
DocumentDbCollectionSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'DocumentDbCollectionSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
nestingSeparator | Nested properties separator. Type: string (or Expression with resultType string). | |
query | Documents query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
DrillSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'DrillSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
DynamicsAXSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'DynamicsAXSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
DynamicsCrmSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'DynamicsCrmSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | FetchXML is a proprietary query language that is used in Microsoft Dynamics CRM (online & on-premises). Type: string (or Expression with resultType string). |
DynamicsSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'DynamicsSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | FetchXML is a proprietary query language that is used in Microsoft Dynamics (online & on-premises). Type: string (or Expression with resultType string). |
EloquaSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'EloquaSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
ExcelSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ExcelSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
storeSettings | Excel store settings. | StoreReadSettings |
FileSystemSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'FileSystemSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). |
GoogleAdWordsSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'GoogleAdWordsSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
GoogleBigQuerySource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'GoogleBigQuerySource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
GoogleBigQueryV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | 'GoogleBigQueryV2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
GreenplumSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'GreenplumSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
HBaseSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'HBaseSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
HdfsSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'HdfsSource' (required) |
distcpSettings | Specifies Distcp-related settings. | DistcpSettings |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). |
HiveSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'HiveSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
HttpSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'HttpSource' (required) |
httpRequestTimeout | Specifies the timeout for a HTTP client to get HTTP response from HTTP server. The default value is equivalent to System.Net.HttpWebRequest.Timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
HubspotSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'HubspotSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
ImpalaSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ImpalaSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
InformixSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'InformixSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
JiraSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'JiraSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
JsonSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'JsonSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
formatSettings | Json format settings. | JsonReadSettings |
storeSettings | Json store settings. | StoreReadSettings |
JsonReadSettings
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
type | The read setting type. | string (required) |
{customized property} |
LakeHouseTableSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'LakeHouseTableSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
timestampAsOf | Query an older snapshot by timestamp. Type: string (or Expression with resultType string). | |
versionAsOf | Query an older snapshot by version. Type: integer (or Expression with resultType integer). |
MagentoSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MagentoSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
MariaDBSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MariaDBSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
MarketoSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MarketoSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
MicrosoftAccessSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MicrosoftAccessSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). |
MongoDbAtlasSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MongoDbAtlasSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
batchSize | Specifies the number of documents to return in each batch of the response from MongoDB Atlas instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer). | |
cursorMethods | Cursor methods for Mongodb query | MongoDbCursorMethodsProperties |
filter | Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
MongoDbSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MongoDbSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Should be a SQL-92 query expression. Type: string (or Expression with resultType string). |
MongoDbV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MongoDbV2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
batchSize | Specifies the number of documents to return in each batch of the response from MongoDB instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer). | |
cursorMethods | Cursor methods for Mongodb query | MongoDbCursorMethodsProperties |
filter | Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
MySqlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'MySqlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
NetezzaSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'NetezzaSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
partitionOption | The partition mechanism that will be used for Netezza read in parallel. Possible values include: "None", "DataSlice", "DynamicRange". | |
partitionSettings | The settings that will be leveraged for Netezza source partitioning. | NetezzaPartitionSettings |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
NetezzaPartitionSettings
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). |
ODataSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ODataSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
query | OData query. For example, "$top=1". Type: string (or Expression with resultType string). |
OdbcSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'OdbcSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Office365Source
Name | Description | Value |
---|---|---|
type | Copy source type. | 'Office365Source' (required) |
allowedGroups | The groups containing all the users. Type: array of strings (or Expression with resultType array of strings). | |
dateFilterColumn | The Column to apply the {paramref name="StartTime"/} and {paramref name="EndTime"/}. Type: string (or Expression with resultType string). | |
endTime | End time of the requested range for this dataset. Type: string (or Expression with resultType string). | |
outputColumns | The columns to be read out from the Office 365 table. Type: array of objects (or Expression with resultType array of objects). itemType: OutputColumn. Example: [ { "name": "Id" }, { "name": "CreatedDateTime" } ] | |
startTime | Start time of the requested range for this dataset. Type: string (or Expression with resultType string). | |
userScopeFilterUri | The user scope uri. Type: string (or Expression with resultType string). |
OracleServiceCloudSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'OracleServiceCloudSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
OracleSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'OracleSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
oracleReaderQuery | Oracle reader query. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Oracle read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". | |
partitionSettings | The settings that will be leveraged for Oracle source partitioning. | OraclePartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
OraclePartitionSettings
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionNames | Names of the physical partitions of Oracle table. | |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). |
OrcSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'OrcSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
storeSettings | ORC store settings. | StoreReadSettings |
ParquetSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ParquetSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
formatSettings | Parquet format settings. | ParquetReadSettings |
storeSettings | Parquet store settings. | StoreReadSettings |
ParquetReadSettings
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
type | The read setting type. | string (required) |
{customized property} |
PaypalSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'PaypalSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
PhoenixSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'PhoenixSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
PostgreSqlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'PostgreSqlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
PostgreSqlV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | 'PostgreSqlV2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
PrestoSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'PrestoSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
QuickBooksSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'QuickBooksSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
RelationalSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'RelationalSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). |
ResponsysSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ResponsysSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
RestSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'RestSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: key value pairs (value should be string type). | |
additionalHeaders | The additional HTTP headers in the request to the RESTful API. Type: string (or Expression with resultType string). | |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
paginationRules | The pagination rules to compose next page requests. Type: string (or Expression with resultType string). | |
requestBody | The HTTP request body to the RESTful API if requestMethod is POST. Type: string (or Expression with resultType string). | |
requestInterval | The time to await before sending next page request. | |
requestMethod | The HTTP method used to call the RESTful API. The default is GET. Type: string (or Expression with resultType string). |
SalesforceMarketingCloudSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SalesforceMarketingCloudSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
SalesforceServiceCloudSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SalesforceServiceCloudSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
readBehavior | The read behavior for the operation. Default is Query. Allowed values: Query/QueryAll. Type: string (or Expression with resultType string). |
SalesforceServiceCloudV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SalesforceServiceCloudV2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
includeDeletedObjects | This property control whether query result contains Deleted objects. Default is false. Type: boolean (or Expression with resultType boolean). | |
SOQLQuery | Database query. Type: string (or Expression with resultType string). |
SalesforceSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SalesforceSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
readBehavior | The read behavior for the operation. Default is Query. Allowed values: Query/QueryAll. Type: string (or Expression with resultType string). |
SalesforceV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SalesforceV2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
includeDeletedObjects | This property control whether query result contains Deleted objects. Default is false. Type: boolean (or Expression with resultType boolean). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
SOQLQuery | Database query. Type: string (or Expression with resultType string). |
SapBwSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapBwSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | MDX query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
SapCloudForCustomerSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapCloudForCustomerSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
query | SAP Cloud for Customer OData query. For example, "$top=1". Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
SapEccSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapEccSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
query | SAP ECC OData query. For example, "$top=1". Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
SapHanaSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapHanaSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
packetSize | The packet size of data read from SAP HANA. Type: integer(or Expression with resultType integer). | |
partitionOption | The partition mechanism that will be used for SAP HANA read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "SapHanaDynamicRange". | |
partitionSettings | The settings that will be leveraged for SAP HANA source partitioning. | SapHanaPartitionSettings |
query | SAP HANA Sql query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
SapHanaPartitionSettings
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). |
SapOdpSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapOdpSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
extractionMode | The extraction mode. Allowed value include: Full, Delta and Recovery. The default value is Full. Type: string (or Expression with resultType string). | |
projection | Specifies the columns to be selected from source data. Type: array of objects(projection) (or Expression with resultType array of objects). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
selection | Specifies the selection conditions from source data. Type: array of objects(selection) (or Expression with resultType array of objects). | |
subscriberProcess | The subscriber process to manage the delta process. Type: string (or Expression with resultType string). |
SapOpenHubSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapOpenHubSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
baseRequestId | The ID of request for delta loading. Once it is set, only data with requestId larger than the value of this property will be retrieved. The default value is 0. Type: integer (or Expression with resultType integer ). | |
customRfcReadTableFunctionModule | Specifies the custom RFC function module that will be used to read data from SAP Table. Type: string (or Expression with resultType string). | |
excludeLastRequest | Whether to exclude the records of the last request. The default value is true. Type: boolean (or Expression with resultType boolean). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sapDataColumnDelimiter | The single character that will be used as delimiter passed to SAP RFC as well as splitting the output data retrieved. Type: string (or Expression with resultType string). |
SapTableSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SapTableSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
batchSize | Specifies the maximum number of rows that will be retrieved at a time when retrieving data from SAP Table. Type: integer (or Expression with resultType integer). | |
customRfcReadTableFunctionModule | Specifies the custom RFC function module that will be used to read data from SAP Table. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for SAP table read in parallel. Possible values include: "None", "PartitionOnInt", "PartitionOnCalendarYear", "PartitionOnCalendarMonth", "PartitionOnCalendarDate", "PartitionOnTime". | |
partitionSettings | The settings that will be leveraged for SAP table source partitioning. | SapTablePartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
rfcTableFields | The fields of the SAP table that will be retrieved. For example, column0, column1. Type: string (or Expression with resultType string). | |
rfcTableOptions | The options for the filtering of the SAP Table. For example, COLUMN0 EQ SOME VALUE. Type: string (or Expression with resultType string). | |
rowCount | The number of rows to be retrieved. Type: integer(or Expression with resultType integer). | |
rowSkips | The number of rows that will be skipped. Type: integer (or Expression with resultType integer). | |
sapDataColumnDelimiter | The single character that will be used as delimiter passed to SAP RFC as well as splitting the output data retrieved. Type: string (or Expression with resultType string). |
SapTablePartitionSettings
Name | Description | Value |
---|---|---|
maxPartitionsNumber | The maximum value of partitions the table will be split into. Type: integer (or Expression with resultType string). | |
partitionColumnName | The name of the column that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). |
ServiceNowSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ServiceNowSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
ServiceNowV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ServiceNowV2Source' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
expression | Expression to filter data from source. | ExpressionV2 |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
ExpressionV2
Name | Description | Value |
---|---|---|
operands | List of nested expressions. | ExpressionV2[] |
operator | Expression operator value Type: string. | string |
type | Type of expressions supported by the system. Type: string. | 'Binary' 'Constant' 'Field' 'Unary' |
value | Value for Constant/Field Type: string. | string |
SharePointOnlineListSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SharePointOnlineListSource' (required) |
httpRequestTimeout | The wait time to get a response from SharePoint Online. Default value is 5 minutes (00:05:00). Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
query | The OData query to filter the data in SharePoint Online list. For example, "$top=1". Type: string (or Expression with resultType string). |
ShopifySource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ShopifySource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
SnowflakeSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SnowflakeSource' (required) |
exportSettings | Snowflake export settings. | SnowflakeExportCopyCommand (required) |
query | Snowflake Sql query. Type: string (or Expression with resultType string). |
SnowflakeExportCopyCommand
Name | Description | Value |
---|---|---|
additionalCopyOptions | Additional copy options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": "'HH24:MI:SS.FF'" } | SnowflakeExportCopyCommandAdditionalCopyOptions |
additionalFormatOptions | Additional format options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalFormatOptions": { "OVERWRITE": "TRUE", "MAX_FILE_SIZE": "'FALSE'" } | SnowflakeExportCopyCommandAdditionalFormatOptions |
type | The export setting type. | string (required) |
{customized property} |
SnowflakeExportCopyCommandAdditionalCopyOptions
Name | Description | Value |
---|---|---|
{customized property} |
SnowflakeExportCopyCommandAdditionalFormatOptions
Name | Description | Value |
---|---|---|
{customized property} |
SnowflakeV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SnowflakeV2Source' (required) |
exportSettings | Snowflake export settings. | SnowflakeExportCopyCommand (required) |
query | Snowflake Sql query. Type: string (or Expression with resultType string). |
SparkSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SparkSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
SqlDWSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SqlDWSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | SQL Data Warehouse reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Data Warehouse source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". Type: object (or Expression with resultType object), itemType: StoredProcedureParameter. |
SqlMISource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SqlMISource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
produceAdditionalTypes | Which additional types to produce. | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a Azure SQL Managed Instance source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". |
SqlServerSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SqlServerSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
produceAdditionalTypes | Which additional types to produce. | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". |
SqlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SqlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". |
SquareSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SquareSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
SybaseSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'SybaseSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
TeradataSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'TeradataSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
partitionOption | The partition mechanism that will be used for teradata read in parallel. Possible values include: "None", "Hash", "DynamicRange". | |
partitionSettings | The settings that will be leveraged for teradata source partitioning. | TeradataPartitionSettings |
query | Teradata query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
TeradataPartitionSettings
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column that will be used for proceeding range or hash partitioning. Type: string (or Expression with resultType string). | |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). |
VerticaSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'VerticaSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
WarehouseSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'WarehouseSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the Microsoft Fabric Warehouse source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | Microsoft Fabric Warehouse reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a Microsoft Fabric Warehouse source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". Type: object (or Expression with resultType object), itemType: StoredProcedureParameter. |
WebSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'WebSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). |
XeroSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'XeroSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
XmlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'XmlSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
formatSettings | Xml format settings. | XmlReadSettings |
storeSettings | Xml store settings. | StoreReadSettings |
XmlReadSettings
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
detectDataType | Indicates whether type detection is enabled when reading the xml files. Type: boolean (or Expression with resultType boolean). | |
namespacePrefixes | Namespace uri to prefix mappings to override the prefixes in column names when namespace is enabled, if no prefix is defined for a namespace uri, the prefix of xml element/attribute name in the xml data file will be used. Example: "{"http://www.example.com/xml":"prefix"}" Type: object (or Expression with resultType object). | |
namespaces | Indicates whether namespace is enabled when reading the xml files. Type: boolean (or Expression with resultType boolean). | |
type | The read setting type. | string (required) |
validationMode | Indicates what validation method is used when reading the xml files. Allowed values: 'none', 'xsd', or 'dtd'. Type: string (or Expression with resultType string). | |
{customized property} |
ZohoSource
Name | Description | Value |
---|---|---|
type | Copy source type. | 'ZohoSource' (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
StagingSettings
Name | Description | Value |
---|---|---|
enableCompression | Specifies whether to use compression when copying data via an interim staging. Default value is false. Type: boolean (or Expression with resultType boolean). | |
linkedServiceName | Staging linked service reference. | LinkedServiceReference (required) |
path | The path to storage for storing the interim data. Type: string (or Expression with resultType string). | |
{customized property} |
CustomActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Custom' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Custom activity properties. | CustomActivityTypeProperties (required) |
CustomActivityTypeProperties
Name | Description | Value |
---|---|---|
autoUserSpecification | Elevation level and scope for the user, default is nonadmin task. Type: string (or Expression with resultType double). | |
command | Command for custom activity Type: string (or Expression with resultType string). | |
extendedProperties | User defined property bag. There is no restriction on the keys or values that can be used. The user specified custom activity has the full responsibility to consume and interpret the content defined. | CustomActivityTypePropertiesExtendedProperties |
folderPath | Folder path for resource files Type: string (or Expression with resultType string). | |
referenceObjects | Reference objects | CustomActivityReferenceObject |
resourceLinkedService | Resource linked service reference. | LinkedServiceReference |
retentionTimeInDays | The retention time for the files submitted for custom activity. Type: double (or Expression with resultType double). |
CustomActivityTypePropertiesExtendedProperties
Name | Description | Value |
---|---|---|
{customized property} |
CustomActivityReferenceObject
Name | Description | Value |
---|---|---|
datasets | Dataset references. | DatasetReference[] |
linkedServices | Linked service references. | LinkedServiceReference[] |
DatabricksNotebookActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'DatabricksNotebook' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Databricks Notebook activity properties. | DatabricksNotebookActivityTypeProperties (required) |
DatabricksNotebookActivityTypeProperties
Name | Description | Value |
---|---|---|
baseParameters | Base parameters to be used for each run of this job.If the notebook takes a parameter that is not specified, the default value from the notebook will be used. | DatabricksNotebookActivityTypePropertiesBaseParamete... |
libraries | A list of libraries to be installed on the cluster that will execute the job. | DatabricksNotebookActivityTypePropertiesLibrariesIte...[] |
notebookPath | The absolute path of the notebook to be run in the Databricks Workspace. This path must begin with a slash. Type: string (or Expression with resultType string). |
DatabricksNotebookActivityTypePropertiesBaseParamete...
Name | Description | Value |
---|---|---|
{customized property} |
DatabricksNotebookActivityTypePropertiesLibrariesIte...
Name | Description | Value |
---|---|---|
{customized property} |
DatabricksSparkJarActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'DatabricksSparkJar' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Databricks SparkJar activity properties. | DatabricksSparkJarActivityTypeProperties (required) |
DatabricksSparkJarActivityTypeProperties
Name | Description | Value |
---|---|---|
libraries | A list of libraries to be installed on the cluster that will execute the job. | DatabricksSparkJarActivityTypePropertiesLibrariesIte...[] |
mainClassName | The full name of the class containing the main method to be executed. This class must be contained in a JAR provided as a library. Type: string (or Expression with resultType string). | |
parameters | Parameters that will be passed to the main method. | any[] |
DatabricksSparkJarActivityTypePropertiesLibrariesIte...
Name | Description | Value |
---|---|---|
{customized property} |
DatabricksSparkPythonActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'DatabricksSparkPython' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Databricks SparkPython activity properties. | DatabricksSparkPythonActivityTypeProperties (required) |
DatabricksSparkPythonActivityTypeProperties
Name | Description | Value |
---|---|---|
libraries | A list of libraries to be installed on the cluster that will execute the job. | DatabricksSparkPythonActivityTypePropertiesLibraries...[] |
parameters | Command line parameters that will be passed to the Python file. | any[] |
pythonFile | The URI of the Python file to be executed. DBFS paths are supported. Type: string (or Expression with resultType string). |
DatabricksSparkPythonActivityTypePropertiesLibraries...
Name | Description | Value |
---|---|---|
{customized property} |
DataLakeAnalyticsUsqlActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'DataLakeAnalyticsU-SQL' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Data Lake Analytics U-SQL activity properties. | DataLakeAnalyticsUsqlActivityTypeProperties (required) |
DataLakeAnalyticsUsqlActivityTypeProperties
Name | Description | Value |
---|---|---|
compilationMode | Compilation mode of U-SQL. Must be one of these values : Semantic, Full and SingleBox. Type: string (or Expression with resultType string). | |
degreeOfParallelism | The maximum number of nodes simultaneously used to run the job. Default value is 1. Type: integer (or Expression with resultType integer), minimum: 1. | |
parameters | Parameters for U-SQL job request. | DataLakeAnalyticsUsqlActivityTypePropertiesParameter... |
priority | Determines which jobs out of all that are queued should be selected to run first. The lower the number, the higher the priority. Default value is 1000. Type: integer (or Expression with resultType integer), minimum: 1. | |
runtimeVersion | Runtime version of the U-SQL engine to use. Type: string (or Expression with resultType string). | |
scriptLinkedService | Script linked service reference. | LinkedServiceReference (required) |
scriptPath | Case-sensitive path to folder that contains the U-SQL script. Type: string (or Expression with resultType string). |
DataLakeAnalyticsUsqlActivityTypePropertiesParameter...
Name | Description | Value |
---|---|---|
{customized property} |
DeleteActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Delete' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Delete activity properties. | DeleteActivityTypeProperties (required) |
DeleteActivityTypeProperties
Name | Description | Value |
---|---|---|
dataset | Delete activity dataset reference. | DatasetReference (required) |
enableLogging | Whether to record detailed logs of delete-activity execution. Default value is false. Type: boolean (or Expression with resultType boolean). | |
logStorageSettings | Log storage settings customer need to provide when enableLogging is true. | LogStorageSettings |
maxConcurrentConnections | The max concurrent connections to connect data source at the same time. | int Constraints: Min value = 1 |
recursive | If true, files or sub-folders under current folder path will be deleted recursively. Default is false. Type: boolean (or Expression with resultType boolean). | |
storeSettings | Delete activity store settings. | StoreReadSettings |
ExecuteDataFlowActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'ExecuteDataFlow' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute data flow activity properties. | ExecuteDataFlowActivityTypeProperties (required) |
ExecuteDataFlowActivityTypeProperties
Name | Description | Value |
---|---|---|
compute | Compute properties for data flow activity. | ExecuteDataFlowActivityTypePropertiesCompute |
continueOnError | Continue on error setting used for data flow execution. Enables processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean) | |
dataFlow | Data flow reference. | DataFlowReference (required) |
integrationRuntime | The integration runtime reference. | IntegrationRuntimeReference |
runConcurrently | Concurrent run setting used for data flow execution. Allows sinks with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean) | |
sourceStagingConcurrency | Specify number of parallel staging for sources applicable to the sink. Type: integer (or Expression with resultType integer) | |
staging | Staging info for execute data flow activity. | DataFlowStagingInfo |
traceLevel | Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string) |
ExecuteDataFlowActivityTypePropertiesCompute
Name | Description | Value |
---|---|---|
computeType | Compute type of the cluster which will execute data flow job. Possible values include: 'General', 'MemoryOptimized', 'ComputeOptimized'. Type: string (or Expression with resultType string) | |
coreCount | Core count of the cluster which will execute data flow job. Supported values are: 8, 16, 32, 48, 80, 144 and 272. Type: integer (or Expression with resultType integer) |
DataFlowReference
Name | Description | Value |
---|---|---|
datasetParameters | Reference data flow parameters from dataset. | |
parameters | Data flow parameters | ParameterValueSpecification |
referenceName | Reference data flow name. | string (required) |
type | Data flow reference type. | 'DataFlowReference' (required) |
{customized property} |
IntegrationRuntimeReference
Name | Description | Value |
---|---|---|
parameters | Arguments for integration runtime. | ParameterValueSpecification |
referenceName | Reference integration runtime name. | string (required) |
type | Type of integration runtime. | 'IntegrationRuntimeReference' (required) |
DataFlowStagingInfo
Name | Description | Value |
---|---|---|
folderPath | Folder path for staging blob. Type: string (or Expression with resultType string) | |
linkedService | Staging linked service reference. | LinkedServiceReference |
ExecutePipelineActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'ExecutePipeline' (required) |
policy | Execute pipeline activity policy. | ExecutePipelineActivityPolicy |
typeProperties | Execute pipeline activity properties. | ExecutePipelineActivityTypeProperties (required) |
ExecutePipelineActivityPolicy
Name | Description | Value |
---|---|---|
secureInput | When set to true, Input from activity is considered as secure and will not be logged to monitoring. | bool |
{customized property} |
ExecutePipelineActivityTypeProperties
Name | Description | Value |
---|---|---|
parameters | Pipeline parameters. | ParameterValueSpecification |
pipeline | Pipeline reference. | PipelineReference (required) |
waitOnCompletion | Defines whether activity execution will wait for the dependent pipeline execution to finish. Default is false. | bool |
PipelineReference
Name | Description | Value |
---|---|---|
name | Reference name. | string |
referenceName | Reference pipeline name. | string (required) |
type | Pipeline reference type. | 'PipelineReference' (required) |
ExecuteSsisPackageActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'ExecuteSSISPackage' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute SSIS package activity properties. | ExecuteSsisPackageActivityTypeProperties (required) |
ExecuteSsisPackageActivityTypeProperties
Name | Description | Value |
---|---|---|
connectVia | The integration runtime reference. | IntegrationRuntimeReference (required) |
environmentPath | The environment path to execute the SSIS package. Type: string (or Expression with resultType string). | |
executionCredential | The package execution credential. | SsisExecutionCredential |
loggingLevel | The logging level of SSIS package execution. Type: string (or Expression with resultType string). | |
logLocation | SSIS package execution log location. | SsisLogLocation |
packageConnectionManagers | The package level connection managers to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesPackageConne... |
packageLocation | SSIS package location. | SsisPackageLocation (required) |
packageParameters | The package level parameters to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesPackageParam... |
projectConnectionManagers | The project level connection managers to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesProjectConne... |
projectParameters | The project level parameters to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesProjectParam... |
propertyOverrides | The property overrides to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesPropertyOver... |
runtime | Specifies the runtime to execute SSIS package. The value should be "x86" or "x64". Type: string (or Expression with resultType string). |
SsisExecutionCredential
Name | Description | Value |
---|---|---|
domain | Domain for windows authentication. Type: string (or Expression with resultType string). | |
password | Password for windows authentication. | SecureString (required) |
userName | UseName for windows authentication. Type: string (or Expression with resultType string). |
SecureString
Name | Description | Value |
---|---|---|
type | Type of the secret. | string (required) |
value | Value of secure string. | string (required) |
SsisLogLocation
Name | Description | Value |
---|---|---|
logPath | The SSIS package execution log path. Type: string (or Expression with resultType string). | |
type | The type of SSIS log location. | 'File' (required) |
typeProperties | SSIS package execution log location properties. | SsisLogLocationTypeProperties (required) |
SsisLogLocationTypeProperties
Name | Description | Value |
---|---|---|
accessCredential | The package execution log access credential. | SsisAccessCredential |
logRefreshInterval | Specifies the interval to refresh log. The default interval is 5 minutes. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
SsisAccessCredential
Name | Description | Value |
---|---|---|
domain | Domain for windows authentication. Type: string (or Expression with resultType string). | |
password | Password for windows authentication. | SecretBase (required) |
userName | UseName for windows authentication. Type: string (or Expression with resultType string). |
SecretBase
Name | Description | Value |
---|---|---|
type | Set the object type | AzureKeyVaultSecret SecureString (required) |
AzureKeyVaultSecretReference
Name | Description | Value |
---|---|---|
secretName | The name of the secret in Azure Key Vault. Type: string (or Expression with resultType string). | |
secretVersion | The version of the secret in Azure Key Vault. The default value is the latest version of the secret. Type: string (or Expression with resultType string). | |
store | The Azure Key Vault linked service reference. | LinkedServiceReference (required) |
type | Type of the secret. | string (required) |
ExecuteSsisPackageActivityTypePropertiesPackageConne...
Name | Description | Value |
---|---|---|
{customized property} | SsisConnectionManager |
SsisConnectionManager
Name | Description | Value |
---|---|---|
{customized property} | SsisExecutionParameter |
SsisExecutionParameter
Name | Description | Value |
---|---|---|
value | SSIS package execution parameter value. Type: string (or Expression with resultType string). |
SsisPackageLocation
Name | Description | Value |
---|---|---|
packagePath | The SSIS package path. Type: string (or Expression with resultType string). | |
type | The type of SSIS package location. | 'File' 'InlinePackage' 'PackageStore' 'SSISDB' |
typeProperties | SSIS package location properties. | SsisPackageLocationTypeProperties |
SsisPackageLocationTypeProperties
Name | Description | Value |
---|---|---|
accessCredential | The package access credential. | SsisAccessCredential |
childPackages | The embedded child package list. | SsisChildPackage[] |
configurationAccessCredential | The configuration file access credential. | SsisAccessCredential |
configurationPath | The configuration file of the package execution. Type: string (or Expression with resultType string). | |
packageContent | The embedded package content. Type: string (or Expression with resultType string). | |
packageLastModifiedDate | The embedded package last modified date. | string |
packageName | The package name. | string |
packagePassword | Password of the package. | SecretBase |
SsisChildPackage
Name | Description | Value |
---|---|---|
packageContent | Content for embedded child package. Type: string (or Expression with resultType string). | |
packageLastModifiedDate | Last modified date for embedded child package. | string |
packageName | Name for embedded child package. | string |
packagePath | Path for embedded child package. Type: string (or Expression with resultType string). |
ExecuteSsisPackageActivityTypePropertiesPackageParam...
Name | Description | Value |
---|---|---|
{customized property} | SsisExecutionParameter |
ExecuteSsisPackageActivityTypePropertiesProjectConne...
Name | Description | Value |
---|---|---|
{customized property} | SsisConnectionManager |
ExecuteSsisPackageActivityTypePropertiesProjectParam...
Name | Description | Value |
---|---|---|
{customized property} | SsisExecutionParameter |
ExecuteSsisPackageActivityTypePropertiesPropertyOver...
Name | Description | Value |
---|---|---|
{customized property} | SsisPropertyOverride |
SsisPropertyOverride
Name | Description | Value |
---|---|---|
isSensitive | Whether SSIS package property override value is sensitive data. Value will be encrypted in SSISDB if it is true | bool |
value | SSIS package property override value. Type: string (or Expression with resultType string). |
ExecuteWranglingDataflowActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'ExecuteWranglingDataflow' (required) |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute power query activity properties. | ExecutePowerQueryActivityTypeProperties (required) |
ExecutePowerQueryActivityTypeProperties
Name | Description | Value |
---|---|---|
compute | Compute properties for data flow activity. | ExecuteDataFlowActivityTypePropertiesCompute |
continueOnError | Continue on error setting used for data flow execution. Enables processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean) | |
dataFlow | Data flow reference. | DataFlowReference (required) |
integrationRuntime | The integration runtime reference. | IntegrationRuntimeReference |
queries | List of mapping for Power Query mashup query to sink dataset(s). | PowerQuerySinkMapping[] |
runConcurrently | Concurrent run setting used for data flow execution. Allows sinks with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean) | |
sinks | (Deprecated. Please use Queries). List of Power Query activity sinks mapped to a queryName. | ExecutePowerQueryActivityTypePropertiesSinks |
sourceStagingConcurrency | Specify number of parallel staging for sources applicable to the sink. Type: integer (or Expression with resultType integer) | |
staging | Staging info for execute data flow activity. | DataFlowStagingInfo |
traceLevel | Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string) |
PowerQuerySinkMapping
Name | Description | Value |
---|---|---|
dataflowSinks | List of sinks mapped to Power Query mashup query. | PowerQuerySink[] |
queryName | Name of the query in Power Query mashup document. | string |
PowerQuerySink
Name | Description | Value |
---|---|---|
dataset | Dataset reference. | DatasetReference |
description | Transformation description. | string |
flowlet | Flowlet Reference | DataFlowReference |
linkedService | Linked service reference. | LinkedServiceReference |
name | Transformation name. | string (required) |
rejectedDataLinkedService | Rejected data linked service reference. | LinkedServiceReference |
schemaLinkedService | Schema linked service reference. | LinkedServiceReference |
script | sink script. | string |
ExecutePowerQueryActivityTypePropertiesSinks
Name | Description | Value |
---|---|---|
{customized property} | PowerQuerySink |
FailActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Fail' (required) |
typeProperties | Fail activity properties. | FailActivityTypeProperties (required) |
FailActivityTypeProperties
Name | Description | Value |
---|---|---|
errorCode | The error code that categorizes the error type of the Fail activity. It can be dynamic content that's evaluated to a non empty/blank string at runtime. Type: string (or Expression with resultType string). | |
message | The error message that surfaced in the Fail activity. It can be dynamic content that's evaluated to a non empty/blank string at runtime. Type: string (or Expression with resultType string). |
FilterActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Filter' (required) |
typeProperties | Filter activity properties. | FilterActivityTypeProperties (required) |
FilterActivityTypeProperties
Name | Description | Value |
---|---|---|
condition | Condition to be used for filtering the input. | Expression (required) |
items | Input array on which filter should be applied. | Expression (required) |
Expression
Name | Description | Value |
---|---|---|
type | Expression type. | 'Expression' (required) |
value | Expression value. | string (required) |
ForEachActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'ForEach' (required) |
typeProperties | ForEach activity properties. | ForEachActivityTypeProperties (required) |
ForEachActivityTypeProperties
Name | Description | Value |
---|---|---|
activities | List of activities to execute . | Activity[] (required) |
batchCount | Batch count to be used for controlling the number of parallel execution (when isSequential is set to false). | int Constraints: Max value = 50 |
isSequential | Should the loop be executed in sequence or in parallel (max 50) | bool |
items | Collection to iterate. | Expression (required) |
GetMetadataActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'GetMetadata' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | GetMetadata activity properties. | GetMetadataActivityTypeProperties (required) |
GetMetadataActivityTypeProperties
Name | Description | Value |
---|---|---|
dataset | GetMetadata activity dataset reference. | DatasetReference (required) |
fieldList | Fields of metadata to get from dataset. | any[] |
formatSettings | GetMetadata activity format settings. | FormatReadSettings |
storeSettings | GetMetadata activity store settings. | StoreReadSettings |
FormatReadSettings
Name | Description | Value |
---|---|---|
type | Set the object type | BinaryReadSettings DelimitedTextReadSettings JsonReadSettings ParquetReadSettings XmlReadSettings (required) |
HDInsightHiveActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'HDInsightHive' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight Hive activity properties. | HDInsightHiveActivityTypeProperties (required) |
HDInsightHiveActivityTypeProperties
Name | Description | Value |
---|---|---|
arguments | User specified arguments to HDInsightActivity. | any[] |
defines | Allows user to specify defines for Hive job request. | HDInsightHiveActivityTypePropertiesDefines |
getDebugInfo | Debug info option. | 'Always' 'Failure' 'None' |
queryTimeout | Query timeout value (in minutes). Effective when the HDInsight cluster is with ESP (Enterprise Security Package) | int |
scriptLinkedService | Script linked service reference. | LinkedServiceReference |
scriptPath | Script path. Type: string (or Expression with resultType string). | |
storageLinkedServices | Storage linked service references. | LinkedServiceReference[] |
variables | User specified arguments under hivevar namespace. | HDInsightHiveActivityTypePropertiesVariables |
HDInsightHiveActivityTypePropertiesDefines
Name | Description | Value |
---|---|---|
{customized property} |
HDInsightHiveActivityTypePropertiesVariables
Name | Description | Value |
---|---|---|
{customized property} |
HDInsightMapReduceActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'HDInsightMapReduce' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight MapReduce activity properties. | HDInsightMapReduceActivityTypeProperties (required) |
HDInsightMapReduceActivityTypeProperties
Name | Description | Value |
---|---|---|
arguments | User specified arguments to HDInsightActivity. | any[] |
className | Class name. Type: string (or Expression with resultType string). | |
defines | Allows user to specify defines for the MapReduce job request. | HDInsightMapReduceActivityTypePropertiesDefines |
getDebugInfo | Debug info option. | 'Always' 'Failure' 'None' |
jarFilePath | Jar path. Type: string (or Expression with resultType string). | |
jarLibs | Jar libs. | any[] |
jarLinkedService | Jar linked service reference. | LinkedServiceReference |
storageLinkedServices | Storage linked service references. | LinkedServiceReference[] |
HDInsightMapReduceActivityTypePropertiesDefines
Name | Description | Value |
---|---|---|
{customized property} |
HDInsightPigActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'HDInsightPig' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight Pig activity properties. | HDInsightPigActivityTypeProperties (required) |
HDInsightPigActivityTypeProperties
Name | Description | Value |
---|---|---|
arguments | User specified arguments to HDInsightActivity. Type: array (or Expression with resultType array). | |
defines | Allows user to specify defines for Pig job request. | HDInsightPigActivityTypePropertiesDefines |
getDebugInfo | Debug info option. | 'Always' 'Failure' 'None' |
scriptLinkedService | Script linked service reference. | LinkedServiceReference |
scriptPath | Script path. Type: string (or Expression with resultType string). | |
storageLinkedServices | Storage linked service references. | LinkedServiceReference[] |
HDInsightPigActivityTypePropertiesDefines
Name | Description | Value |
---|---|---|
{customized property} |
HDInsightSparkActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'HDInsightSpark' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight spark activity properties. | HDInsightSparkActivityTypeProperties (required) |
HDInsightSparkActivityTypeProperties
Name | Description | Value |
---|---|---|
arguments | The user-specified arguments to HDInsightSparkActivity. | any[] |
className | The application's Java/Spark main class. | string |
entryFilePath | The relative path to the root folder of the code/package to be executed. Type: string (or Expression with resultType string). | |
getDebugInfo | Debug info option. | 'Always' 'Failure' 'None' |
proxyUser | The user to impersonate that will execute the job. Type: string (or Expression with resultType string). | |
rootPath | The root path in 'sparkJobLinkedService' for all the job’s files. Type: string (or Expression with resultType string). | |
sparkConfig | Spark configuration property. | HDInsightSparkActivityTypePropertiesSparkConfig |
sparkJobLinkedService | The storage linked service for uploading the entry file and dependencies, and for receiving logs. | LinkedServiceReference |
HDInsightSparkActivityTypePropertiesSparkConfig
Name | Description | Value |
---|---|---|
{customized property} |
HDInsightStreamingActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'HDInsightStreaming' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight streaming activity properties. | HDInsightStreamingActivityTypeProperties (required) |
HDInsightStreamingActivityTypeProperties
Name | Description | Value |
---|---|---|
arguments | User specified arguments to HDInsightActivity. | any[] |
combiner | Combiner executable name. Type: string (or Expression with resultType string). | |
commandEnvironment | Command line environment values. | any[] |
defines | Allows user to specify defines for streaming job request. | HDInsightStreamingActivityTypePropertiesDefines |
fileLinkedService | Linked service reference where the files are located. | LinkedServiceReference |
filePaths | Paths to streaming job files. Can be directories. | any[] (required) |
getDebugInfo | Debug info option. | 'Always' 'Failure' 'None' |
input | Input blob path. Type: string (or Expression with resultType string). | |
mapper | Mapper executable name. Type: string (or Expression with resultType string). | |
output | Output blob path. Type: string (or Expression with resultType string). | |
reducer | Reducer executable name. Type: string (or Expression with resultType string). | |
storageLinkedServices | Storage linked service references. | LinkedServiceReference[] |
HDInsightStreamingActivityTypePropertiesDefines
Name | Description | Value |
---|---|---|
{customized property} |
IfConditionActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'IfCondition' (required) |
typeProperties | IfCondition activity properties. | IfConditionActivityTypeProperties (required) |
IfConditionActivityTypeProperties
Name | Description | Value |
---|---|---|
expression | An expression that would evaluate to Boolean. This is used to determine the block of activities (ifTrueActivities or ifFalseActivities) that will be executed. | Expression (required) |
ifFalseActivities | List of activities to execute if expression is evaluated to false. This is an optional property and if not provided, the activity will exit without any action. | Activity[] |
ifTrueActivities | List of activities to execute if expression is evaluated to true. This is an optional property and if not provided, the activity will exit without any action. | Activity[] |
LookupActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Lookup' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Lookup activity properties. | LookupActivityTypeProperties (required) |
LookupActivityTypeProperties
Name | Description | Value |
---|---|---|
dataset | Lookup activity dataset reference. | DatasetReference (required) |
firstRowOnly | Whether to return first row or all rows. Default value is true. Type: boolean (or Expression with resultType boolean). | |
source | Dataset-specific source properties, same as copy activity source. | CopySource (required) |
ScriptActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Script' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Script activity properties. | ScriptActivityTypeProperties (required) |
ScriptActivityTypeProperties
Name | Description | Value |
---|---|---|
logSettings | Log settings of script activity. | ScriptActivityTypePropertiesLogSettings |
scriptBlockExecutionTimeout | ScriptBlock execution timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
scripts | Array of script blocks. Type: array. | ScriptActivityScriptBlock[] |
ScriptActivityTypePropertiesLogSettings
Name | Description | Value |
---|---|---|
logDestination | The destination of logs. Type: string. | 'ActivityOutput' 'ExternalStore' (required) |
logLocationSettings | Log location settings customer needs to provide when enabling log. | LogLocationSettings |
ScriptActivityScriptBlock
Name | Description | Value |
---|---|---|
parameters | Array of script parameters. Type: array. | ScriptActivityParameter[] |
text | The query text. Type: string (or Expression with resultType string). | |
type | The type of the query. Type: string. | 'NonQuery' 'Query' (required) |
ScriptActivityParameter
Name | Description | Value |
---|---|---|
direction | The direction of the parameter. | 'Input' 'InputOutput' 'Output' |
name | The name of the parameter. Type: string (or Expression with resultType string). | |
size | The size of the output direction parameter. | int |
type | The type of the parameter. | 'Boolean' 'DateTime' 'DateTimeOffset' 'Decimal' 'Double' 'Guid' 'Int16' 'Int32' 'Int64' 'Single' 'String' 'Timespan' |
value | The value of the parameter. Type: string (or Expression with resultType string). |
SetVariableActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'SetVariable' (required) |
policy | Activity policy. | SecureInputOutputPolicy |
typeProperties | Set Variable activity properties. | SetVariableActivityTypeProperties (required) |
SecureInputOutputPolicy
Name | Description | Value |
---|---|---|
secureInput | When set to true, Input from activity is considered as secure and will not be logged to monitoring. | bool |
secureOutput | When set to true, Output from activity is considered as secure and will not be logged to monitoring. | bool |
SetVariableActivityTypeProperties
Name | Description | Value |
---|---|---|
setSystemVariable | If set to true, it sets the pipeline run return value. | bool |
value | Value to be set. Could be a static value or Expression. | |
variableName | Name of the variable whose value needs to be set. | string |
SynapseSparkJobDefinitionActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'SparkJob' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute spark job activity properties. | SynapseSparkJobActivityTypeProperties (required) |
SynapseSparkJobActivityTypeProperties
Name | Description | Value |
---|---|---|
args | User specified arguments to SynapseSparkJobDefinitionActivity. | any[] |
className | The fully-qualified identifier or the main class that is in the main definition file, which will override the 'className' of the spark job definition you provide. Type: string (or Expression with resultType string). | |
conf | Spark configuration properties, which will override the 'conf' of the spark job definition you provide. | |
configurationType | The type of the spark config. | 'Artifact' 'Customized' 'Default' |
driverSize | Number of core and memory to be used for driver allocated in the specified Spark pool for the job, which will be used for overriding 'driverCores' and 'driverMemory' of the spark job definition you provide. Type: string (or Expression with resultType string). | |
executorSize | Number of core and memory to be used for executors allocated in the specified Spark pool for the job, which will be used for overriding 'executorCores' and 'executorMemory' of the spark job definition you provide. Type: string (or Expression with resultType string). | |
file | The main file used for the job, which will override the 'file' of the spark job definition you provide. Type: string (or Expression with resultType string). | |
files | (Deprecated. Please use pythonCodeReference and filesV2) Additional files used for reference in the main definition file, which will override the 'files' of the spark job definition you provide. | any[] |
filesV2 | Additional files used for reference in the main definition file, which will override the 'jars' and 'files' of the spark job definition you provide. | any[] |
numExecutors | Number of executors to launch for this job, which will override the 'numExecutors' of the spark job definition you provide. Type: integer (or Expression with resultType integer). | |
pythonCodeReference | Additional python code files used for reference in the main definition file, which will override the 'pyFiles' of the spark job definition you provide. | any[] |
scanFolder | Scanning subfolders from the root folder of the main definition file, these files will be added as reference files. The folders named 'jars', 'pyFiles', 'files' or 'archives' will be scanned, and the folders name are case sensitive. Type: boolean (or Expression with resultType boolean). | |
sparkConfig | Spark configuration property. | SynapseSparkJobActivityTypePropertiesSparkConfig |
sparkJob | Synapse spark job reference. | SynapseSparkJobReference (required) |
targetBigDataPool | The name of the big data pool which will be used to execute the spark batch job, which will override the 'targetBigDataPool' of the spark job definition you provide. | BigDataPoolParametrizationReference |
targetSparkConfiguration | The spark configuration of the spark job. | SparkConfigurationParametrizationReference |
SynapseSparkJobActivityTypePropertiesSparkConfig
Name | Description | Value |
---|---|---|
{customized property} |
SynapseSparkJobReference
Name | Description | Value |
---|---|---|
referenceName | Reference spark job name. Expression with resultType string. | |
type | Synapse spark job reference type. | 'SparkJobDefinitionReference' (required) |
BigDataPoolParametrizationReference
Name | Description | Value |
---|---|---|
referenceName | Reference big data pool name. Type: string (or Expression with resultType string). | |
type | Big data pool reference type. | 'BigDataPoolReference' (required) |
SparkConfigurationParametrizationReference
Name | Description | Value |
---|---|---|
referenceName | Reference spark configuration name. Type: string (or Expression with resultType string). | |
type | Spark configuration reference type. | 'SparkConfigurationReference' (required) |
SqlServerStoredProcedureActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'SqlServerStoredProcedure' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | SQL stored procedure activity properties. | SqlServerStoredProcedureActivityTypeProperties (required) |
SqlServerStoredProcedureActivityTypeProperties
Name | Description | Value |
---|---|---|
storedProcedureName | Stored procedure name. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". |
SwitchActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Switch' (required) |
typeProperties | Switch activity properties. | SwitchActivityTypeProperties (required) |
SwitchActivityTypeProperties
Name | Description | Value |
---|---|---|
cases | List of cases that correspond to expected values of the 'on' property. This is an optional property and if not provided, the activity will execute activities provided in defaultActivities. | SwitchCase[] |
defaultActivities | List of activities to execute if no case condition is satisfied. This is an optional property and if not provided, the activity will exit without any action. | Activity[] |
on | An expression that would evaluate to a string or integer. This is used to determine the block of activities in cases that will be executed. | Expression (required) |
SwitchCase
Name | Description | Value |
---|---|---|
activities | List of activities to execute for satisfied case condition. | Activity[] |
value | Expected value that satisfies the expression result of the 'on' property. | string |
SynapseNotebookActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'SynapseNotebook' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute Synapse notebook activity properties. | SynapseNotebookActivityTypeProperties (required) |
SynapseNotebookActivityTypeProperties
Name | Description | Value |
---|---|---|
conf | Spark configuration properties, which will override the 'conf' of the notebook you provide. | |
configurationType | The type of the spark config. | 'Artifact' 'Customized' 'Default' |
driverSize | Number of core and memory to be used for driver allocated in the specified Spark pool for the session, which will be used for overriding 'driverCores' and 'driverMemory' of the notebook you provide. Type: string (or Expression with resultType string). | |
executorSize | Number of core and memory to be used for executors allocated in the specified Spark pool for the session, which will be used for overriding 'executorCores' and 'executorMemory' of the notebook you provide. Type: string (or Expression with resultType string). | |
notebook | Synapse notebook reference. | SynapseNotebookReference (required) |
numExecutors | Number of executors to launch for this session, which will override the 'numExecutors' of the notebook you provide. Type: integer (or Expression with resultType integer). | |
parameters | Notebook parameters. | SynapseNotebookActivityTypePropertiesParameters |
sparkConfig | Spark configuration property. | SynapseNotebookActivityTypePropertiesSparkConfig |
sparkPool | The name of the big data pool which will be used to execute the notebook. | BigDataPoolParametrizationReference |
targetSparkConfiguration | The spark configuration of the spark job. | SparkConfigurationParametrizationReference |
SynapseNotebookReference
Name | Description | Value |
---|---|---|
referenceName | Reference notebook name. Type: string (or Expression with resultType string). | |
type | Synapse notebook reference type. | 'NotebookReference' (required) |
SynapseNotebookActivityTypePropertiesParameters
Name | Description | Value |
---|---|---|
{customized property} | NotebookParameter |
NotebookParameter
Name | Description | Value |
---|---|---|
type | Notebook parameter type. | 'bool' 'float' 'int' 'string' |
value | Notebook parameter value. Type: string (or Expression with resultType string). |
SynapseNotebookActivityTypePropertiesSparkConfig
Name | Description | Value |
---|---|---|
{customized property} |
UntilActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Until' (required) |
typeProperties | Until activity properties. | UntilActivityTypeProperties (required) |
UntilActivityTypeProperties
Name | Description | Value |
---|---|---|
activities | List of activities to execute. | Activity[] (required) |
expression | An expression that would evaluate to Boolean. The loop will continue until this expression evaluates to true | Expression (required) |
timeout | Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
ValidationActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Validation' (required) |
typeProperties | Validation activity properties. | ValidationActivityTypeProperties (required) |
ValidationActivityTypeProperties
Name | Description | Value |
---|---|---|
childItems | Can be used if dataset points to a folder. If set to true, the folder must have at least one file. If set to false, the folder must be empty. Type: boolean (or Expression with resultType boolean). | |
dataset | Validation activity dataset reference. | DatasetReference (required) |
minimumSize | Can be used if dataset points to a file. The file must be greater than or equal in size to the value specified. Type: integer (or Expression with resultType integer). | |
sleep | A delay in seconds between validation attempts. If no value is specified, 10 seconds will be used as the default. Type: integer (or Expression with resultType integer). | |
timeout | Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
WaitActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'Wait' (required) |
typeProperties | Wait activity properties. | WaitActivityTypeProperties (required) |
WaitActivityTypeProperties
Name | Description | Value |
---|---|---|
waitTimeInSeconds | Duration in seconds. Type: integer (or Expression with resultType integer). |
WebActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'WebActivity' (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Web activity properties. | WebActivityTypeProperties (required) |
WebActivityTypeProperties
Name | Description | Value |
---|---|---|
authentication | Authentication method used for calling the endpoint. | WebActivityAuthentication |
body | Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). | |
connectVia | The integration runtime reference. | IntegrationRuntimeReference |
datasets | List of datasets passed to web endpoint. | DatasetReference[] |
disableCertValidation | When set to true, Certificate validation will be disabled. | bool |
headers | Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). | WebActivityTypePropertiesHeaders |
httpRequestTimeout | Timeout for the HTTP request to get a response. Format is in TimeSpan (hh:mm:ss). This value is the timeout to get a response, not the activity timeout. The default value is 00:01:00 (1 minute). The range is from 1 to 10 minutes | |
linkedServices | List of linked services passed to web endpoint. | LinkedServiceReference[] |
method | Rest API method for target endpoint. | 'DELETE' 'GET' 'POST' 'PUT' (required) |
turnOffAsync | Option to disable invoking HTTP GET on location given in response header of a HTTP 202 Response. If set true, it stops invoking HTTP GET on http location given in response header. If set false then continues to invoke HTTP GET call on location given in http response headers. | bool |
url | Web activity target endpoint and path. Type: string (or Expression with resultType string). |
WebActivityAuthentication
Name | Description | Value |
---|---|---|
credential | The credential reference containing authentication information. | CredentialReference |
password | Password for the PFX file or basic authentication / Secret when used for ServicePrincipal | SecretBase |
pfx | Base64-encoded contents of a PFX file or Certificate when used for ServicePrincipal | SecretBase |
resource | Resource for which Azure Auth token will be requested when using MSI Authentication. Type: string (or Expression with resultType string). | |
type | Web activity authentication (Basic/ClientCertificate/MSI/ServicePrincipal) | string |
username | Web activity authentication user name for basic authentication or ClientID when used for ServicePrincipal. Type: string (or Expression with resultType string). | |
userTenant | TenantId for which Azure Auth token will be requested when using ServicePrincipal Authentication. Type: string (or Expression with resultType string). |
CredentialReference
Name | Description | Value |
---|---|---|
referenceName | Reference credential name. | string (required) |
type | Credential reference type. | 'CredentialReference' (required) |
{customized property} |
WebActivityTypePropertiesHeaders
Name | Description | Value |
---|---|---|
{customized property} | string |
WebHookActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | 'WebHook' (required) |
policy | Activity policy. | SecureInputOutputPolicy |
typeProperties | WebHook activity properties. | WebHookActivityTypeProperties (required) |
WebHookActivityTypeProperties
Name | Description | Value |
---|---|---|
authentication | Authentication method used for calling the endpoint. | WebActivityAuthentication |
body | Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). | |
headers | Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). | WebHookActivityTypePropertiesHeaders |
method | Rest API method for target endpoint. | 'POST' (required) |
reportStatusOnCallBack | When set to true, statusCode, output and error in callback request body will be consumed by activity. The activity can be marked as failed by setting statusCode >= 400 in callback request. Default is false. Type: boolean (or Expression with resultType boolean). | |
timeout | The timeout within which the webhook should be called back. If there is no value specified, it defaults to 10 minutes. Type: string. Pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | string |
url | WebHook activity target endpoint and path. Type: string (or Expression with resultType string). |
WebHookActivityTypePropertiesHeaders
Name | Description | Value |
---|---|---|
{customized property} | string |
PipelineFolder
Name | Description | Value |
---|---|---|
name | The name of the folder that this Pipeline is in. | string |
ParameterDefinitionSpecification
Name | Description | Value |
---|---|---|
{customized property} | ParameterSpecification |
ParameterSpecification
Name | Description | Value |
---|---|---|
defaultValue | Default value of parameter. | |
type | Parameter type. | 'Array' 'Bool' 'Float' 'Int' 'Object' 'SecureString' 'String' (required) |
PipelinePolicy
Name | Description | Value |
---|---|---|
elapsedTimeMetric | Pipeline ElapsedTime Metric Policy. | PipelineElapsedTimeMetricPolicy |
PipelineElapsedTimeMetricPolicy
Name | Description | Value |
---|---|---|
duration | TimeSpan value, after which an Azure Monitoring Metric is fired. |
PipelineRunDimensions
Name | Description | Value |
---|---|---|
{customized property} |
VariableDefinitionSpecification
Name | Description | Value |
---|---|---|
{customized property} | VariableSpecification |
VariableSpecification
Name | Description | Value |
---|---|---|
defaultValue | Default value of variable. | |
type | Variable type. | 'Array' 'Bool' 'String' (required) |
Quickstart templates
The following quickstart templates deploy this resource type.
Template | Description |
---|---|
More is possible with Azure Data Factory - One click to try Azure Data Factory |
This template creates a data factory pipeline for a copy activity from Azure Blob into another Azure Blob |
Create a Data Factory, copy from Salesforce to Blob |
This template uses Azure Data Factory to perform data movement from Salesforce to Azure Blobs |
Bulk copy using Azure Data Factory |
This template creates a V2 data factory that copies data in bulk an Azure SQL database to an Azure SQL data warehouse. |
Create a V2 data factory |
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to another folder in the storage. |
Create a V2 data factory (MySQL) |
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to a table in an Azure Database for MySQL. |
Create a V2 data factory (PostGre) |
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to a table in an Azure Database for PostgreSQL. |
Create a V2 data factory (SQL) |
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to a table in an Azure SQL database. |
Create a V2 data factory (SQL On-prem) |
This template creates a V2 data factory that copies data from an on-premises SQL Server to an Azure blob storage |
Create a V2 data factory (Spark) |
This template creates a V2 data factory that copies data from a folder in an Azure Blob Storage to another folder in the storage. |
Terraform (AzAPI provider) resource definition
The factories/pipelines resource type can be deployed with operations that target:
- Resource groups
For a list of changed properties in each API version, see change log.
Resource format
To create a Microsoft.DataFactory/factories/pipelines resource, add the following Terraform to your template.
resource "azapi_resource" "symbolicname" {
type = "Microsoft.DataFactory/factories/pipelines@2018-06-01"
name = "string"
parent_id = "string"
body = jsonencode({
properties = {
activities = [
{
dependsOn = [
{
activity = "string"
dependencyConditions = [
"string"
]
}
]
description = "string"
name = "string"
onInactiveMarkAs = "string"
state = "string"
userProperties = [
{
name = "string"
}
]
type = "string"
// For remaining properties, see Activity objects
}
]
annotations = [ object ]
concurrency = int
description = "string"
folder = {
name = "string"
}
parameters = {
{customized property} = {
type = "string"
}
}
policy = {
elapsedTimeMetric = {}
}
runDimensions = {}
variables = {
{customized property} = {
type = "string"
}
}
}
})
}
Activity objects
Set the type property to specify the type of object.
For AppendVariable, use:
type = "AppendVariable"
typeProperties = {
variableName = "string"
}
For AzureDataExplorerCommand, use:
type = "AzureDataExplorerCommand"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {}
For AzureFunctionActivity, use:
type = "AzureFunctionActivity"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
headers = {
{customized property} = "string"
}
method = "string"
}
For AzureMLBatchExecution, use:
type = "AzureMLBatchExecution"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
globalParameters = {}
webServiceInputs = {
{customized property} = {
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
}
}
webServiceOutputs = {
{customized property} = {
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
}
}
}
For AzureMLExecutePipeline, use:
type = "AzureMLExecutePipeline"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {}
For AzureMLUpdateResource, use:
type = "AzureMLUpdateResource"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
trainedModelLinkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
}
For Copy, use:
type = "Copy"
inputs = [
{
parameters = {}
referenceName = "string"
type = "DatasetReference"
}
]
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
outputs = [
{
parameters = {}
referenceName = "string"
type = "DatasetReference"
}
]
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
logSettings = {
copyActivityLogSettings = {}
logLocationSettings = {
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
}
}
logStorageSettings = {
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
}
preserve = [ object ]
preserveRules = [ object ]
redirectIncompatibleRowSettings = {}
sink = {
type = "string"
// For remaining properties, see CopySink objects
}
skipErrorFile = {}
source = {
type = "string"
// For remaining properties, see CopySource objects
}
stagingSettings = {
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
}
}
For Custom, use:
type = "Custom"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
extendedProperties = {}
referenceObjects = {
datasets = [
{
parameters = {}
referenceName = "string"
type = "DatasetReference"
}
]
linkedServices = [
{
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
]
}
resourceLinkedService = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
}
For DatabricksNotebook, use:
type = "DatabricksNotebook"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
baseParameters = {}
libraries = [
{
}
]
}
For DatabricksSparkJar, use:
type = "DatabricksSparkJar"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
libraries = [
{
}
]
parameters = [ object ]
}
For DatabricksSparkPython, use:
type = "DatabricksSparkPython"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
libraries = [
{
}
]
parameters = [ object ]
}
For DataLakeAnalyticsU-SQL, use:
type = "DataLakeAnalyticsU-SQL"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
parameters = {}
scriptLinkedService = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
}
For Delete, use:
type = "Delete"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
dataset = {
parameters = {}
referenceName = "string"
type = "DatasetReference"
}
logStorageSettings = {
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
}
maxConcurrentConnections = int
storeSettings = {
type = "string"
// For remaining properties, see StoreReadSettings objects
}
}
For ExecuteDataFlow, use:
type = "ExecuteDataFlow"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
compute = {}
dataFlow = {
parameters = {}
referenceName = "string"
type = "DataFlowReference"
}
integrationRuntime = {
parameters = {}
referenceName = "string"
type = "IntegrationRuntimeReference"
}
staging = {
linkedService = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
}
}
For ExecutePipeline, use:
type = "ExecutePipeline"
policy = {
secureInput = bool
}
typeProperties = {
parameters = {}
pipeline = {
name = "string"
referenceName = "string"
type = "PipelineReference"
}
waitOnCompletion = bool
}
For ExecuteSSISPackage, use:
type = "ExecuteSSISPackage"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
connectVia = {
parameters = {}
referenceName = "string"
type = "IntegrationRuntimeReference"
}
executionCredential = {
password = {
type = "string"
value = "string"
}
}
logLocation = {
type = "File"
typeProperties = {
accessCredential = {
password = {
type = "string"
// For remaining properties, see SecretBase objects
}
}
}
}
packageConnectionManagers = {
{customized property} = {
{customized property} = {}
}
}
packageLocation = {
type = "string"
typeProperties = {
accessCredential = {
password = {
type = "string"
// For remaining properties, see SecretBase objects
}
}
childPackages = [
{
packageLastModifiedDate = "string"
packageName = "string"
}
]
configurationAccessCredential = {
password = {
type = "string"
// For remaining properties, see SecretBase objects
}
}
packageLastModifiedDate = "string"
packageName = "string"
packagePassword = {
type = "string"
// For remaining properties, see SecretBase objects
}
}
}
packageParameters = {
{customized property} = {}
}
projectConnectionManagers = {
{customized property} = {
{customized property} = {}
}
}
projectParameters = {
{customized property} = {}
}
propertyOverrides = {
{customized property} = {
isSensitive = bool
}
}
}
For ExecuteWranglingDataflow, use:
type = "ExecuteWranglingDataflow"
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
compute = {}
dataFlow = {
parameters = {}
referenceName = "string"
type = "DataFlowReference"
}
integrationRuntime = {
parameters = {}
referenceName = "string"
type = "IntegrationRuntimeReference"
}
queries = [
{
dataflowSinks = [
{
dataset = {
parameters = {}
referenceName = "string"
type = "DatasetReference"
}
description = "string"
flowlet = {
parameters = {}
referenceName = "string"
type = "DataFlowReference"
}
linkedService = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
name = "string"
rejectedDataLinkedService = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
schemaLinkedService = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
script = "string"
}
]
queryName = "string"
}
]
sinks = {
{customized property} = {
dataset = {
parameters = {}
referenceName = "string"
type = "DatasetReference"
}
description = "string"
flowlet = {
parameters = {}
referenceName = "string"
type = "DataFlowReference"
}
linkedService = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
name = "string"
rejectedDataLinkedService = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
schemaLinkedService = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
script = "string"
}
}
staging = {
linkedService = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
}
}
For Fail, use:
type = "Fail"
typeProperties = {}
For Filter, use:
type = "Filter"
typeProperties = {
condition = {
type = "Expression"
value = "string"
}
items = {
type = "Expression"
value = "string"
}
}
For ForEach, use:
type = "ForEach"
typeProperties = {
activities = [
{
dependsOn = [
{
activity = "string"
dependencyConditions = [
"string"
]
}
]
description = "string"
name = "string"
onInactiveMarkAs = "string"
state = "string"
userProperties = [
{
name = "string"
}
]
type = "string"
// For remaining properties, see Activity objects
}
]
batchCount = int
isSequential = bool
items = {
type = "Expression"
value = "string"
}
}
For GetMetadata, use:
type = "GetMetadata"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
dataset = {
parameters = {}
referenceName = "string"
type = "DatasetReference"
}
fieldList = [ object ]
formatSettings = {
type = "string"
// For remaining properties, see FormatReadSettings objects
}
storeSettings = {
type = "string"
// For remaining properties, see StoreReadSettings objects
}
}
For HDInsightHive, use:
type = "HDInsightHive"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
arguments = [ object ]
defines = {}
getDebugInfo = "string"
queryTimeout = int
scriptLinkedService = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
storageLinkedServices = [
{
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
]
variables = {}
}
For HDInsightMapReduce, use:
type = "HDInsightMapReduce"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
arguments = [ object ]
defines = {}
getDebugInfo = "string"
jarLibs = [ object ]
jarLinkedService = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
storageLinkedServices = [
{
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
]
}
For HDInsightPig, use:
type = "HDInsightPig"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
defines = {}
getDebugInfo = "string"
scriptLinkedService = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
storageLinkedServices = [
{
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
]
}
For HDInsightSpark, use:
type = "HDInsightSpark"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
arguments = [ object ]
className = "string"
getDebugInfo = "string"
sparkConfig = {}
sparkJobLinkedService = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
}
For HDInsightStreaming, use:
type = "HDInsightStreaming"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
arguments = [ object ]
commandEnvironment = [ object ]
defines = {}
fileLinkedService = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
filePaths = [ object ]
getDebugInfo = "string"
storageLinkedServices = [
{
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
]
}
For IfCondition, use:
type = "IfCondition"
typeProperties = {
expression = {
type = "Expression"
value = "string"
}
ifFalseActivities = [
{
dependsOn = [
{
activity = "string"
dependencyConditions = [
"string"
]
}
]
description = "string"
name = "string"
onInactiveMarkAs = "string"
state = "string"
userProperties = [
{
name = "string"
}
]
type = "string"
// For remaining properties, see Activity objects
}
]
ifTrueActivities = [
{
dependsOn = [
{
activity = "string"
dependencyConditions = [
"string"
]
}
]
description = "string"
name = "string"
onInactiveMarkAs = "string"
state = "string"
userProperties = [
{
name = "string"
}
]
type = "string"
// For remaining properties, see Activity objects
}
]
}
For Lookup, use:
type = "Lookup"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
dataset = {
parameters = {}
referenceName = "string"
type = "DatasetReference"
}
source = {
type = "string"
// For remaining properties, see CopySource objects
}
}
For Script, use:
type = "Script"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
logSettings = {
logDestination = "string"
logLocationSettings = {
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
}
}
scripts = [
{
parameters = [
{
direction = "string"
size = int
type = "string"
}
]
type = "string"
}
]
}
For SetVariable, use:
type = "SetVariable"
policy = {
secureInput = bool
secureOutput = bool
}
typeProperties = {
setSystemVariable = bool
variableName = "string"
}
For SparkJob, use:
type = "SparkJob"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
args = [ object ]
configurationType = "string"
files = [ object ]
filesV2 = [ object ]
pythonCodeReference = [ object ]
sparkConfig = {}
sparkJob = {
type = "SparkJobDefinitionReference"
}
targetBigDataPool = {
type = "BigDataPoolReference"
}
targetSparkConfiguration = {
type = "SparkConfigurationReference"
}
}
For SqlServerStoredProcedure, use:
type = "SqlServerStoredProcedure"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {}
For Switch, use:
type = "Switch"
typeProperties = {
cases = [
{
activities = [
{
dependsOn = [
{
activity = "string"
dependencyConditions = [
"string"
]
}
]
description = "string"
name = "string"
onInactiveMarkAs = "string"
state = "string"
userProperties = [
{
name = "string"
}
]
type = "string"
// For remaining properties, see Activity objects
}
]
value = "string"
}
]
defaultActivities = [
{
dependsOn = [
{
activity = "string"
dependencyConditions = [
"string"
]
}
]
description = "string"
name = "string"
onInactiveMarkAs = "string"
state = "string"
userProperties = [
{
name = "string"
}
]
type = "string"
// For remaining properties, see Activity objects
}
]
on = {
type = "Expression"
value = "string"
}
}
For SynapseNotebook, use:
type = "SynapseNotebook"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
configurationType = "string"
notebook = {
type = "NotebookReference"
}
parameters = {
{customized property} = {
type = "string"
}
}
sparkConfig = {}
sparkPool = {
type = "BigDataPoolReference"
}
targetSparkConfiguration = {
type = "SparkConfigurationReference"
}
}
For Until, use:
type = "Until"
typeProperties = {
activities = [
{
dependsOn = [
{
activity = "string"
dependencyConditions = [
"string"
]
}
]
description = "string"
name = "string"
onInactiveMarkAs = "string"
state = "string"
userProperties = [
{
name = "string"
}
]
type = "string"
// For remaining properties, see Activity objects
}
]
expression = {
type = "Expression"
value = "string"
}
}
For Validation, use:
type = "Validation"
typeProperties = {
dataset = {
parameters = {}
referenceName = "string"
type = "DatasetReference"
}
}
For Wait, use:
type = "Wait"
typeProperties = {}
For WebActivity, use:
type = "WebActivity"
linkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
policy = {
retryIntervalInSeconds = int
secureInput = bool
secureOutput = bool
}
typeProperties = {
authentication = {
credential = {
referenceName = "string"
type = "CredentialReference"
}
password = {
type = "string"
// For remaining properties, see SecretBase objects
}
pfx = {
type = "string"
// For remaining properties, see SecretBase objects
}
type = "string"
}
connectVia = {
parameters = {}
referenceName = "string"
type = "IntegrationRuntimeReference"
}
datasets = [
{
parameters = {}
referenceName = "string"
type = "DatasetReference"
}
]
disableCertValidation = bool
headers = {
{customized property} = "string"
}
linkedServices = [
{
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
]
method = "string"
turnOffAsync = bool
}
For WebHook, use:
type = "WebHook"
policy = {
secureInput = bool
secureOutput = bool
}
typeProperties = {
authentication = {
credential = {
referenceName = "string"
type = "CredentialReference"
}
password = {
type = "string"
// For remaining properties, see SecretBase objects
}
pfx = {
type = "string"
// For remaining properties, see SecretBase objects
}
type = "string"
}
headers = {
{customized property} = "string"
}
method = "POST"
timeout = "string"
}
CopySink objects
Set the type property to specify the type of object.
For AvroSink, use:
type = "AvroSink"
formatSettings = {
recordName = "string"
recordNamespace = "string"
type = "string"
}
storeSettings = {
metadata = [
{
}
]
type = "string"
// For remaining properties, see StoreWriteSettings objects
}
For AzureBlobFSSink, use:
type = "AzureBlobFSSink"
metadata = [
{
}
]
For AzureDatabricksDeltaLakeSink, use:
type = "AzureDatabricksDeltaLakeSink"
importSettings = {
type = "string"
}
For AzureDataExplorerSink, use:
type = "AzureDataExplorerSink"
For AzureDataLakeStoreSink, use:
type = "AzureDataLakeStoreSink"
For AzureMySqlSink, use:
type = "AzureMySqlSink"
For AzurePostgreSqlSink, use:
type = "AzurePostgreSqlSink"
For AzureQueueSink, use:
type = "AzureQueueSink"
For AzureSearchIndexSink, use:
type = "AzureSearchIndexSink"
writeBehavior = "string"
For AzureSqlSink, use:
type = "AzureSqlSink"
upsertSettings = {}
For AzureTableSink, use:
type = "AzureTableSink"
For BinarySink, use:
type = "BinarySink"
storeSettings = {
metadata = [
{
}
]
type = "string"
// For remaining properties, see StoreWriteSettings objects
}
For BlobSink, use:
type = "BlobSink"
metadata = [
{
}
]
For CommonDataServiceForAppsSink, use:
type = "CommonDataServiceForAppsSink"
writeBehavior = "Upsert"
For CosmosDbMongoDbApiSink, use:
type = "CosmosDbMongoDbApiSink"
For CosmosDbSqlApiSink, use:
type = "CosmosDbSqlApiSink"
For DelimitedTextSink, use:
type = "DelimitedTextSink"
formatSettings = {
type = "string"
}
storeSettings = {
metadata = [
{
}
]
type = "string"
// For remaining properties, see StoreWriteSettings objects
}
For DocumentDbCollectionSink, use:
type = "DocumentDbCollectionSink"
For DynamicsCrmSink, use:
type = "DynamicsCrmSink"
writeBehavior = "Upsert"
For DynamicsSink, use:
type = "DynamicsSink"
writeBehavior = "Upsert"
For FileSystemSink, use:
type = "FileSystemSink"
For InformixSink, use:
type = "InformixSink"
For JsonSink, use:
type = "JsonSink"
formatSettings = {
type = "string"
}
storeSettings = {
metadata = [
{
}
]
type = "string"
// For remaining properties, see StoreWriteSettings objects
}
For LakeHouseTableSink, use:
type = "LakeHouseTableSink"
For MicrosoftAccessSink, use:
type = "MicrosoftAccessSink"
For MongoDbAtlasSink, use:
type = "MongoDbAtlasSink"
For MongoDbV2Sink, use:
type = "MongoDbV2Sink"
For OdbcSink, use:
type = "OdbcSink"
For OracleSink, use:
type = "OracleSink"
For OrcSink, use:
type = "OrcSink"
formatSettings = {
type = "string"
}
storeSettings = {
metadata = [
{
}
]
type = "string"
// For remaining properties, see StoreWriteSettings objects
}
For ParquetSink, use:
type = "ParquetSink"
formatSettings = {
type = "string"
}
storeSettings = {
metadata = [
{
}
]
type = "string"
// For remaining properties, see StoreWriteSettings objects
}
For RestSink, use:
type = "RestSink"
For SalesforceServiceCloudSink, use:
type = "SalesforceServiceCloudSink"
writeBehavior = "string"
For SalesforceServiceCloudV2Sink, use:
type = "SalesforceServiceCloudV2Sink"
writeBehavior = "string"
For SalesforceSink, use:
type = "SalesforceSink"
writeBehavior = "string"
For SalesforceV2Sink, use:
type = "SalesforceV2Sink"
writeBehavior = "string"
For SapCloudForCustomerSink, use:
type = "SapCloudForCustomerSink"
writeBehavior = "string"
For SnowflakeSink, use:
type = "SnowflakeSink"
importSettings = {
additionalCopyOptions = {}
additionalFormatOptions = {}
type = "string"
}
For SnowflakeV2Sink, use:
type = "SnowflakeV2Sink"
importSettings = {
additionalCopyOptions = {}
additionalFormatOptions = {}
type = "string"
}
For SqlDWSink, use:
type = "SqlDWSink"
copyCommandSettings = {
additionalOptions = {
{customized property} = "string"
}
defaultValues = [
{
}
]
}
polyBaseSettings = {
rejectType = "string"
}
upsertSettings = {}
For SqlMISink, use:
type = "SqlMISink"
upsertSettings = {}
For SqlServerSink, use:
type = "SqlServerSink"
upsertSettings = {}
For SqlSink, use:
type = "SqlSink"
upsertSettings = {}
For WarehouseSink, use:
type = "WarehouseSink"
copyCommandSettings = {
additionalOptions = {
{customized property} = "string"
}
defaultValues = [
{
}
]
}
StoreWriteSettings objects
Set the type property to specify the type of object.
For AzureBlobFSWriteSettings, use:
type = "AzureBlobFSWriteSettings"
For AzureBlobStorageWriteSettings, use:
type = "AzureBlobStorageWriteSettings"
For AzureDataLakeStoreWriteSettings, use:
type = "AzureDataLakeStoreWriteSettings"
For AzureFileStorageWriteSettings, use:
type = "AzureFileStorageWriteSettings"
For FileServerWriteSettings, use:
type = "FileServerWriteSettings"
For LakeHouseWriteSettings, use:
type = "LakeHouseWriteSettings"
For SftpWriteSettings, use:
type = "SftpWriteSettings"
CopySource objects
Set the type property to specify the type of object.
For AmazonMWSSource, use:
type = "AmazonMWSSource"
For AmazonRdsForOracleSource, use:
type = "AmazonRdsForOracleSource"
partitionSettings = {}
For AmazonRdsForSqlServerSource, use:
type = "AmazonRdsForSqlServerSource"
partitionSettings = {}
For AmazonRedshiftSource, use:
type = "AmazonRedshiftSource"
redshiftUnloadSettings = {
s3LinkedServiceName = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
}
For AvroSource, use:
type = "AvroSource"
storeSettings = {
type = "string"
// For remaining properties, see StoreReadSettings objects
}
For AzureBlobFSSource, use:
type = "AzureBlobFSSource"
For AzureDatabricksDeltaLakeSource, use:
type = "AzureDatabricksDeltaLakeSource"
exportSettings = {
type = "string"
}
For AzureDataExplorerSource, use:
type = "AzureDataExplorerSource"
For AzureDataLakeStoreSource, use:
type = "AzureDataLakeStoreSource"
For AzureMariaDBSource, use:
type = "AzureMariaDBSource"
For AzureMySqlSource, use:
type = "AzureMySqlSource"
For AzurePostgreSqlSource, use:
type = "AzurePostgreSqlSource"
For AzureSqlSource, use:
type = "AzureSqlSource"
partitionSettings = {}
For AzureTableSource, use:
type = "AzureTableSource"
For BinarySource, use:
type = "BinarySource"
formatSettings = {
compressionProperties = {
type = "string"
// For remaining properties, see CompressionReadSettings objects
}
type = "string"
}
storeSettings = {
type = "string"
// For remaining properties, see StoreReadSettings objects
}
For BlobSource, use:
type = "BlobSource"
For CassandraSource, use:
type = "CassandraSource"
consistencyLevel = "string"
For CommonDataServiceForAppsSource, use:
type = "CommonDataServiceForAppsSource"
For ConcurSource, use:
type = "ConcurSource"
For CosmosDbMongoDbApiSource, use:
type = "CosmosDbMongoDbApiSource"
cursorMethods = {}
For CosmosDbSqlApiSource, use:
type = "CosmosDbSqlApiSource"
For CouchbaseSource, use:
type = "CouchbaseSource"
For Db2Source, use:
type = "Db2Source"
For DelimitedTextSource, use:
type = "DelimitedTextSource"
formatSettings = {
compressionProperties = {
type = "string"
// For remaining properties, see CompressionReadSettings objects
}
type = "string"
}
storeSettings = {
type = "string"
// For remaining properties, see StoreReadSettings objects
}
For DocumentDbCollectionSource, use:
type = "DocumentDbCollectionSource"
For DrillSource, use:
type = "DrillSource"
For DynamicsAXSource, use:
type = "DynamicsAXSource"
For DynamicsCrmSource, use:
type = "DynamicsCrmSource"
For DynamicsSource, use:
type = "DynamicsSource"
For EloquaSource, use:
type = "EloquaSource"
For ExcelSource, use:
type = "ExcelSource"
storeSettings = {
type = "string"
// For remaining properties, see StoreReadSettings objects
}
For FileSystemSource, use:
type = "FileSystemSource"
For GoogleAdWordsSource, use:
type = "GoogleAdWordsSource"
For GoogleBigQuerySource, use:
type = "GoogleBigQuerySource"
For GoogleBigQueryV2Source, use:
type = "GoogleBigQueryV2Source"
For GreenplumSource, use:
type = "GreenplumSource"
For HBaseSource, use:
type = "HBaseSource"
For HdfsSource, use:
type = "HdfsSource"
distcpSettings = {}
For HiveSource, use:
type = "HiveSource"
For HttpSource, use:
type = "HttpSource"
For HubspotSource, use:
type = "HubspotSource"
For ImpalaSource, use:
type = "ImpalaSource"
For InformixSource, use:
type = "InformixSource"
For JiraSource, use:
type = "JiraSource"
For JsonSource, use:
type = "JsonSource"
formatSettings = {
compressionProperties = {
type = "string"
// For remaining properties, see CompressionReadSettings objects
}
type = "string"
}
storeSettings = {
type = "string"
// For remaining properties, see StoreReadSettings objects
}
For LakeHouseTableSource, use:
type = "LakeHouseTableSource"
For MagentoSource, use:
type = "MagentoSource"
For MariaDBSource, use:
type = "MariaDBSource"
For MarketoSource, use:
type = "MarketoSource"
For MicrosoftAccessSource, use:
type = "MicrosoftAccessSource"
For MongoDbAtlasSource, use:
type = "MongoDbAtlasSource"
cursorMethods = {}
For MongoDbSource, use:
type = "MongoDbSource"
For MongoDbV2Source, use:
type = "MongoDbV2Source"
cursorMethods = {}
For MySqlSource, use:
type = "MySqlSource"
For NetezzaSource, use:
type = "NetezzaSource"
partitionSettings = {}
For ODataSource, use:
type = "ODataSource"
For OdbcSource, use:
type = "OdbcSource"
For Office365Source, use:
type = "Office365Source"
For OracleServiceCloudSource, use:
type = "OracleServiceCloudSource"
For OracleSource, use:
type = "OracleSource"
partitionSettings = {}
For OrcSource, use:
type = "OrcSource"
storeSettings = {
type = "string"
// For remaining properties, see StoreReadSettings objects
}
For ParquetSource, use:
type = "ParquetSource"
formatSettings = {
compressionProperties = {
type = "string"
// For remaining properties, see CompressionReadSettings objects
}
type = "string"
}
storeSettings = {
type = "string"
// For remaining properties, see StoreReadSettings objects
}
For PaypalSource, use:
type = "PaypalSource"
For PhoenixSource, use:
type = "PhoenixSource"
For PostgreSqlSource, use:
type = "PostgreSqlSource"
For PostgreSqlV2Source, use:
type = "PostgreSqlV2Source"
For PrestoSource, use:
type = "PrestoSource"
For QuickBooksSource, use:
type = "QuickBooksSource"
For RelationalSource, use:
type = "RelationalSource"
For ResponsysSource, use:
type = "ResponsysSource"
For RestSource, use:
type = "RestSource"
For SalesforceMarketingCloudSource, use:
type = "SalesforceMarketingCloudSource"
For SalesforceServiceCloudSource, use:
type = "SalesforceServiceCloudSource"
For SalesforceServiceCloudV2Source, use:
type = "SalesforceServiceCloudV2Source"
For SalesforceSource, use:
type = "SalesforceSource"
For SalesforceV2Source, use:
type = "SalesforceV2Source"
For SapBwSource, use:
type = "SapBwSource"
For SapCloudForCustomerSource, use:
type = "SapCloudForCustomerSource"
For SapEccSource, use:
type = "SapEccSource"
For SapHanaSource, use:
type = "SapHanaSource"
partitionSettings = {}
For SapOdpSource, use:
type = "SapOdpSource"
For SapOpenHubSource, use:
type = "SapOpenHubSource"
For SapTableSource, use:
type = "SapTableSource"
partitionSettings = {}
For ServiceNowSource, use:
type = "ServiceNowSource"
For ServiceNowV2Source, use:
type = "ServiceNowV2Source"
expression = {
operands = [
{
operator = "string"
type = "string"
value = "string"
}
For SharePointOnlineListSource, use:
type = "SharePointOnlineListSource"
For ShopifySource, use:
type = "ShopifySource"
For SnowflakeSource, use:
type = "SnowflakeSource"
exportSettings = {
additionalCopyOptions = {}
additionalFormatOptions = {}
type = "string"
}
For SnowflakeV2Source, use:
type = "SnowflakeV2Source"
exportSettings = {
additionalCopyOptions = {}
additionalFormatOptions = {}
type = "string"
}
For SparkSource, use:
type = "SparkSource"
For SqlDWSource, use:
type = "SqlDWSource"
partitionSettings = {}
For SqlMISource, use:
type = "SqlMISource"
partitionSettings = {}
For SqlServerSource, use:
type = "SqlServerSource"
partitionSettings = {}
For SqlSource, use:
type = "SqlSource"
partitionSettings = {}
For SquareSource, use:
type = "SquareSource"
For SybaseSource, use:
type = "SybaseSource"
For TeradataSource, use:
type = "TeradataSource"
partitionSettings = {}
For VerticaSource, use:
type = "VerticaSource"
For WarehouseSource, use:
type = "WarehouseSource"
partitionSettings = {}
For WebSource, use:
type = "WebSource"
For XeroSource, use:
type = "XeroSource"
For XmlSource, use:
type = "XmlSource"
formatSettings = {
compressionProperties = {
type = "string"
// For remaining properties, see CompressionReadSettings objects
}
type = "string"
}
storeSettings = {
type = "string"
// For remaining properties, see StoreReadSettings objects
}
For ZohoSource, use:
type = "ZohoSource"
StoreReadSettings objects
Set the type property to specify the type of object.
For AmazonS3CompatibleReadSettings, use:
type = "AmazonS3CompatibleReadSettings"
For AmazonS3ReadSettings, use:
type = "AmazonS3ReadSettings"
For AzureBlobFSReadSettings, use:
type = "AzureBlobFSReadSettings"
For AzureBlobStorageReadSettings, use:
type = "AzureBlobStorageReadSettings"
For AzureDataLakeStoreReadSettings, use:
type = "AzureDataLakeStoreReadSettings"
For AzureFileStorageReadSettings, use:
type = "AzureFileStorageReadSettings"
For FileServerReadSettings, use:
type = "FileServerReadSettings"
For FtpReadSettings, use:
type = "FtpReadSettings"
For GoogleCloudStorageReadSettings, use:
type = "GoogleCloudStorageReadSettings"
For HdfsReadSettings, use:
type = "HdfsReadSettings"
distcpSettings = {}
For HttpReadSettings, use:
type = "HttpReadSettings"
For LakeHouseReadSettings, use:
type = "LakeHouseReadSettings"
For OracleCloudStorageReadSettings, use:
type = "OracleCloudStorageReadSettings"
For SftpReadSettings, use:
type = "SftpReadSettings"
CompressionReadSettings objects
Set the type property to specify the type of object.
For TarGZipReadSettings, use:
type = "TarGZipReadSettings"
For TarReadSettings, use:
type = "TarReadSettings"
For ZipDeflateReadSettings, use:
type = "ZipDeflateReadSettings"
SecretBase objects
Set the type property to specify the type of object.
For AzureKeyVaultSecret, use:
type = "AzureKeyVaultSecret"
store = {
parameters = {}
referenceName = "string"
type = "LinkedServiceReference"
}
For SecureString, use:
type = "SecureString"
value = "string"
FormatReadSettings objects
Set the type property to specify the type of object.
For BinaryReadSettings, use:
type = "BinaryReadSettings"
compressionProperties = {
type = "string"
// For remaining properties, see CompressionReadSettings objects
}
For DelimitedTextReadSettings, use:
type = "DelimitedTextReadSettings"
compressionProperties = {
type = "string"
// For remaining properties, see CompressionReadSettings objects
}
For JsonReadSettings, use:
type = "JsonReadSettings"
compressionProperties = {
type = "string"
// For remaining properties, see CompressionReadSettings objects
}
For ParquetReadSettings, use:
type = "ParquetReadSettings"
compressionProperties = {
type = "string"
// For remaining properties, see CompressionReadSettings objects
}
For XmlReadSettings, use:
type = "XmlReadSettings"
compressionProperties = {
type = "string"
// For remaining properties, see CompressionReadSettings objects
}
Property values
factories/pipelines
Name | Description | Value |
---|---|---|
type | The resource type | "Microsoft.DataFactory/factories/pipelines@2018-06-01" |
name | The resource name | string (required) Character limit: 1-260 Valid characters: Can't use: <>*#.%&:\\+?/ or control charactersStart with alphanumeric. |
parent_id | The ID of the resource that is the parent for this resource. | ID for resource of type: factories |
properties | Properties of the pipeline. | Pipeline (required) |
Pipeline
Name | Description | Value |
---|---|---|
activities | List of activities in pipeline. | Activity[] |
annotations | List of tags that can be used for describing the Pipeline. | any[] |
concurrency | The max number of concurrent runs for the pipeline. | int Constraints: Min value = 1 |
description | The description of the pipeline. | string |
folder | The folder that this Pipeline is in. If not specified, Pipeline will appear at the root level. | PipelineFolder |
parameters | List of parameters for pipeline. | ParameterDefinitionSpecification |
policy | Pipeline Policy. | PipelinePolicy |
runDimensions | Dimensions emitted by Pipeline. | PipelineRunDimensions |
variables | List of variables for pipeline. | VariableDefinitionSpecification |
Activity
Name | Description | Value |
---|---|---|
dependsOn | Activity depends on condition. | ActivityDependency[] |
description | Activity description. | string |
name | Activity name. | string (required) |
onInactiveMarkAs | Status result of the activity when the state is set to Inactive. This is an optional property and if not provided when the activity is inactive, the status will be Succeeded by default. | "Failed" "Skipped" "Succeeded" |
state | Activity state. This is an optional property and if not provided, the state will be Active by default. | "Active" "Inactive" |
userProperties | Activity user properties. | UserProperty[] |
type | Set the object type | AppendVariable AzureDataExplorerCommand AzureFunctionActivity AzureMLBatchExecution AzureMLExecutePipeline AzureMLUpdateResource Copy Custom DatabricksNotebook DatabricksSparkJar DatabricksSparkPython DataLakeAnalyticsU-SQL Delete ExecuteDataFlow ExecutePipeline ExecuteSSISPackage ExecuteWranglingDataflow Fail Filter ForEach GetMetadata HDInsightHive HDInsightMapReduce HDInsightPig HDInsightSpark HDInsightStreaming IfCondition Lookup Script SetVariable SparkJob SqlServerStoredProcedure Switch SynapseNotebook Until Validation Wait WebActivity WebHook (required) |
ActivityDependency
Name | Description | Value |
---|---|---|
activity | Activity name. | string (required) |
dependencyConditions | Match-Condition for the dependency. | String array containing any of: "Completed" "Failed" "Skipped" "Succeeded" (required) |
{customized property} |
UserProperty
Name | Description | Value |
---|---|---|
name | User property name. | string (required) |
value | User property value. Type: string (or Expression with resultType string). |
AppendVariableActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "AppendVariable" (required) |
typeProperties | Append Variable activity properties. | AppendVariableActivityTypeProperties (required) |
AppendVariableActivityTypeProperties
Name | Description | Value |
---|---|---|
value | Value to be appended. Type: could be a static value matching type of the variable item or Expression with resultType matching type of the variable item | |
variableName | Name of the variable whose value needs to be appended to. | string |
AzureDataExplorerCommandActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "AzureDataExplorerCommand" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure Data Explorer command activity properties. | AzureDataExplorerCommandActivityTypeProperties (required) |
LinkedServiceReference
Name | Description | Value |
---|---|---|
parameters | Arguments for LinkedService. | ParameterValueSpecification |
referenceName | Reference LinkedService name. | string (required) |
type | Linked service reference type. | "LinkedServiceReference" (required) |
ParameterValueSpecification
Name | Description | Value |
---|---|---|
{customized property} |
ActivityPolicy
Name | Description | Value |
---|---|---|
retry | Maximum ordinary retry attempts. Default is 0. Type: integer (or Expression with resultType integer), minimum: 0. | |
retryIntervalInSeconds | Interval between each retry attempt (in seconds). The default is 30 sec. | int Constraints: Min value = 30 Max value = 86400 |
secureInput | When set to true, Input from activity is considered as secure and will not be logged to monitoring. | bool |
secureOutput | When set to true, Output from activity is considered as secure and will not be logged to monitoring. | bool |
timeout | Specifies the timeout for the activity to run. The default timeout is 7 days. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
{customized property} |
AzureDataExplorerCommandActivityTypeProperties
Name | Description | Value |
---|---|---|
command | A control command, according to the Azure Data Explorer command syntax. Type: string (or Expression with resultType string). | |
commandTimeout | Control command timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9]))..) |
AzureFunctionActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "AzureFunctionActivity" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure Function activity properties. | AzureFunctionActivityTypeProperties (required) |
AzureFunctionActivityTypeProperties
Name | Description | Value |
---|---|---|
body | Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). | |
functionName | Name of the Function that the Azure Function Activity will call. Type: string (or Expression with resultType string) | |
headers | Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). | AzureFunctionActivityTypePropertiesHeaders |
method | Rest API method for target endpoint. | "DELETE" "GET" "HEAD" "OPTIONS" "POST" "PUT" "TRACE" (required) |
AzureFunctionActivityTypePropertiesHeaders
Name | Description | Value |
---|---|---|
{customized property} | string |
AzureMLBatchExecutionActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "AzureMLBatchExecution" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure ML Batch Execution activity properties. | AzureMLBatchExecutionActivityTypeProperties (required) |
AzureMLBatchExecutionActivityTypeProperties
Name | Description | Value |
---|---|---|
globalParameters | Key,Value pairs to be passed to the Azure ML Batch Execution Service endpoint. Keys must match the names of web service parameters defined in the published Azure ML web service. Values will be passed in the GlobalParameters property of the Azure ML batch execution request. | AzureMLBatchExecutionActivityTypePropertiesGlobalPar... |
webServiceInputs | Key,Value pairs, mapping the names of Azure ML endpoint's Web Service Inputs to AzureMLWebServiceFile objects specifying the input Blob locations.. This information will be passed in the WebServiceInputs property of the Azure ML batch execution request. | AzureMLBatchExecutionActivityTypePropertiesWebServic... |
webServiceOutputs | Key,Value pairs, mapping the names of Azure ML endpoint's Web Service Outputs to AzureMLWebServiceFile objects specifying the output Blob locations. This information will be passed in the WebServiceOutputs property of the Azure ML batch execution request. | AzureMLBatchExecutionActivityTypePropertiesWebServic... |
AzureMLBatchExecutionActivityTypePropertiesGlobalPar...
Name | Description | Value |
---|---|---|
{customized property} |
AzureMLBatchExecutionActivityTypePropertiesWebServic...
Name | Description | Value |
---|---|---|
{customized property} | AzureMLWebServiceFile |
AzureMLWebServiceFile
Name | Description | Value |
---|---|---|
filePath | The relative file path, including container name, in the Azure Blob Storage specified by the LinkedService. Type: string (or Expression with resultType string). | |
linkedServiceName | Reference to an Azure Storage LinkedService, where Azure ML WebService Input/Output file located. | LinkedServiceReference (required) |
AzureMLBatchExecutionActivityTypePropertiesWebServic...
Name | Description | Value |
---|---|---|
{customized property} | AzureMLWebServiceFile |
AzureMLExecutePipelineActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "AzureMLExecutePipeline" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure ML Execute Pipeline activity properties. | AzureMLExecutePipelineActivityTypeProperties (required) |
AzureMLExecutePipelineActivityTypeProperties
Name | Description | Value |
---|---|---|
continueOnStepFailure | Whether to continue execution of other steps in the PipelineRun if a step fails. This information will be passed in the continueOnStepFailure property of the published pipeline execution request. Type: boolean (or Expression with resultType boolean). | |
dataPathAssignments | Dictionary used for changing data path assignments without retraining. Values will be passed in the dataPathAssignments property of the published pipeline execution request. Type: object (or Expression with resultType object). | |
experimentName | Run history experiment name of the pipeline run. This information will be passed in the ExperimentName property of the published pipeline execution request. Type: string (or Expression with resultType string). | |
mlParentRunId | The parent Azure ML Service pipeline run id. This information will be passed in the ParentRunId property of the published pipeline execution request. Type: string (or Expression with resultType string). | |
mlPipelineEndpointId | ID of the published Azure ML pipeline endpoint. Type: string (or Expression with resultType string). | |
mlPipelineId | ID of the published Azure ML pipeline. Type: string (or Expression with resultType string). | |
mlPipelineParameters | Key,Value pairs to be passed to the published Azure ML pipeline endpoint. Keys must match the names of pipeline parameters defined in the published pipeline. Values will be passed in the ParameterAssignments property of the published pipeline execution request. Type: object with key value pairs (or Expression with resultType object). | |
version | Version of the published Azure ML pipeline endpoint. Type: string (or Expression with resultType string). |
AzureMLUpdateResourceActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "AzureMLUpdateResource" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Azure ML Update Resource management activity properties. | AzureMLUpdateResourceActivityTypeProperties (required) |
AzureMLUpdateResourceActivityTypeProperties
Name | Description | Value |
---|---|---|
trainedModelFilePath | The relative file path in trainedModelLinkedService to represent the .ilearner file that will be uploaded by the update operation. Type: string (or Expression with resultType string). | |
trainedModelLinkedServiceName | Name of Azure Storage linked service holding the .ilearner file that will be uploaded by the update operation. | LinkedServiceReference (required) |
trainedModelName | Name of the Trained Model module in the Web Service experiment to be updated. Type: string (or Expression with resultType string). |
CopyActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "Copy" (required) |
inputs | List of inputs for the activity. | DatasetReference[] |
linkedServiceName | Linked service reference. | LinkedServiceReference |
outputs | List of outputs for the activity. | DatasetReference[] |
policy | Activity policy. | ActivityPolicy |
typeProperties | Copy activity properties. | CopyActivityTypeProperties (required) |
DatasetReference
Name | Description | Value |
---|---|---|
parameters | Arguments for dataset. | ParameterValueSpecification |
referenceName | Reference dataset name. | string (required) |
type | Dataset reference type. | "DatasetReference" (required) |
CopyActivityTypeProperties
Name | Description | Value |
---|---|---|
dataIntegrationUnits | Maximum number of data integration units that can be used to perform this data movement. Type: integer (or Expression with resultType integer), minimum: 0. | |
enableSkipIncompatibleRow | Whether to skip incompatible row. Default value is false. Type: boolean (or Expression with resultType boolean). | |
enableStaging | Specifies whether to copy data via an interim staging. Default value is false. Type: boolean (or Expression with resultType boolean). | |
logSettings | Log settings customer needs provide when enabling log. | LogSettings |
logStorageSettings | (Deprecated. Please use LogSettings) Log storage settings customer need to provide when enabling session log. | LogStorageSettings |
parallelCopies | Maximum number of concurrent sessions opened on the source or sink to avoid overloading the data store. Type: integer (or Expression with resultType integer), minimum: 0. | |
preserve | Preserve rules. | any[] |
preserveRules | Preserve Rules. | any[] |
redirectIncompatibleRowSettings | Redirect incompatible row settings when EnableSkipIncompatibleRow is true. | RedirectIncompatibleRowSettings |
sink | Copy activity sink. | CopySink (required) |
skipErrorFile | Specify the fault tolerance for data consistency. | SkipErrorFile |
source | Copy activity source. | CopySource (required) |
stagingSettings | Specifies interim staging settings when EnableStaging is true. | StagingSettings |
translator | Copy activity translator. If not specified, tabular translator is used. | |
validateDataConsistency | Whether to enable Data Consistency validation. Type: boolean (or Expression with resultType boolean). |
LogSettings
Name | Description | Value |
---|---|---|
copyActivityLogSettings | Specifies settings for copy activity log. | CopyActivityLogSettings |
enableCopyActivityLog | Specifies whether to enable copy activity log. Type: boolean (or Expression with resultType boolean). | |
logLocationSettings | Log location settings customer needs to provide when enabling log. | LogLocationSettings (required) |
CopyActivityLogSettings
Name | Description | Value |
---|---|---|
enableReliableLogging | Specifies whether to enable reliable logging. Type: boolean (or Expression with resultType boolean). | |
logLevel | Gets or sets the log level, support: Info, Warning. Type: string (or Expression with resultType string). |
LogLocationSettings
Name | Description | Value |
---|---|---|
linkedServiceName | Log storage linked service reference. | LinkedServiceReference (required) |
path | The path to storage for storing detailed logs of activity execution. Type: string (or Expression with resultType string). |
LogStorageSettings
Name | Description | Value |
---|---|---|
enableReliableLogging | Specifies whether to enable reliable logging. Type: boolean (or Expression with resultType boolean). | |
linkedServiceName | Log storage linked service reference. | LinkedServiceReference (required) |
logLevel | Gets or sets the log level, support: Info, Warning. Type: string (or Expression with resultType string). | |
path | The path to storage for storing detailed logs of activity execution. Type: string (or Expression with resultType string). | |
{customized property} |
RedirectIncompatibleRowSettings
Name | Description | Value |
---|---|---|
linkedServiceName | Name of the Azure Storage, Storage SAS, or Azure Data Lake Store linked service used for redirecting incompatible row. Must be specified if redirectIncompatibleRowSettings is specified. Type: string (or Expression with resultType string). | |
path | The path for storing the redirect incompatible row data. Type: string (or Expression with resultType string). | |
{customized property} |
CopySink
Name | Description | Value |
---|---|---|
disableMetricsCollection | If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). | |
maxConcurrentConnections | The maximum concurrent connection count for the sink data store. Type: integer (or Expression with resultType integer). | |
sinkRetryCount | Sink retry count. Type: integer (or Expression with resultType integer). | |
sinkRetryWait | Sink retry wait. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
writeBatchSize | Write batch size. Type: integer (or Expression with resultType integer), minimum: 0. | |
writeBatchTimeout | Write batch timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
type | Set the object type | AvroSink AzureBlobFSSink AzureDatabricksDeltaLakeSink AzureDataExplorerSink AzureDataLakeStoreSink AzureMySqlSink AzurePostgreSqlSink AzureQueueSink AzureSearchIndexSink AzureSqlSink AzureTableSink BinarySink BlobSink CommonDataServiceForAppsSink CosmosDbMongoDbApiSink CosmosDbSqlApiSink DelimitedTextSink DocumentDbCollectionSink DynamicsCrmSink DynamicsSink FileSystemSink InformixSink JsonSink LakeHouseTableSink MicrosoftAccessSink MongoDbAtlasSink MongoDbV2Sink OdbcSink OracleSink OrcSink ParquetSink RestSink SalesforceServiceCloudSink SalesforceServiceCloudV2Sink SalesforceSink SalesforceV2Sink SapCloudForCustomerSink SnowflakeSink SnowflakeV2Sink SqlDWSink SqlMISink SqlServerSink SqlSink WarehouseSink (required) |
AvroSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "AvroSink" (required) |
formatSettings | Avro format settings. | AvroWriteSettings |
storeSettings | Avro store settings. | StoreWriteSettings |
AvroWriteSettings
Name | Description | Value |
---|---|---|
fileNamePrefix | Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). | |
maxRowsPerFile | Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). | |
recordName | Top level record name in write result, which is required in AVRO spec. | string |
recordNamespace | Record namespace in the write result. | string |
type | The write setting type. | string (required) |
{customized property} |
StoreWriteSettings
Name | Description | Value |
---|---|---|
copyBehavior | The type of copy behavior for copy sink. | |
disableMetricsCollection | If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). | |
maxConcurrentConnections | The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). | |
metadata | Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). | MetadataItem[] |
type | Set the object type | AzureBlobFSWriteSettings AzureBlobStorageWriteSettings AzureDataLakeStoreWriteSettings AzureFileStorageWriteSettings FileServerWriteSettings LakeHouseWriteSettings SftpWriteSettings (required) |
MetadataItem
Name | Description | Value |
---|---|---|
name | Metadata item key name. Type: string (or Expression with resultType string). | |
value | Metadata item value. Type: string (or Expression with resultType string). |
AzureBlobFSWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | "AzureBlobFSWriteSettings" (required) |
blockSizeInMB | Indicates the block size(MB) when writing data to blob. Type: integer (or Expression with resultType integer). |
AzureBlobStorageWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | "AzureBlobStorageWriteSettings" (required) |
blockSizeInMB | Indicates the block size(MB) when writing data to blob. Type: integer (or Expression with resultType integer). |
AzureDataLakeStoreWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | "AzureDataLakeStoreWriteSettings" (required) |
expiryDateTime | Specifies the expiry time of the written files. The time is applied to the UTC time zone in the format of "2018-12-01T05:00:00Z". Default value is NULL. Type: string (or Expression with resultType string). |
AzureFileStorageWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | "AzureFileStorageWriteSettings" (required) |
FileServerWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | "FileServerWriteSettings" (required) |
LakeHouseWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | "LakeHouseWriteSettings" (required) |
SftpWriteSettings
Name | Description | Value |
---|---|---|
type | The write setting type. | "SftpWriteSettings" (required) |
operationTimeout | Specifies the timeout for writing each chunk to SFTP server. Default value: 01:00:00 (one hour). Type: string (or Expression with resultType string). | |
useTempFileRename | Upload to temporary file(s) and rename. Disable this option if your SFTP server doesn't support rename operation. Type: boolean (or Expression with resultType boolean). |
AzureBlobFSSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "AzureBlobFSSink" (required) |
copyBehavior | The type of copy behavior for copy sink. Type: string (or Expression with resultType string). | |
metadata | Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). | MetadataItem[] |
AzureDatabricksDeltaLakeSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "AzureDatabricksDeltaLakeSink" (required) |
importSettings | Azure Databricks Delta Lake import settings. | AzureDatabricksDeltaLakeImportCommand |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). |
AzureDatabricksDeltaLakeImportCommand
Name | Description | Value |
---|---|---|
dateFormat | Specify the date format for csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). | |
timestampFormat | Specify the timestamp format for csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). | |
type | The import setting type. | string (required) |
{customized property} |
AzureDataExplorerSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "AzureDataExplorerSink" (required) |
flushImmediately | If set to true, any aggregation will be skipped. Default is false. Type: boolean. | |
ingestionMappingAsJson | An explicit column mapping description provided in a json format. Type: string. | |
ingestionMappingName | A name of a pre-created csv mapping that was defined on the target Kusto table. Type: string. |
AzureDataLakeStoreSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "AzureDataLakeStoreSink" (required) |
copyBehavior | The type of copy behavior for copy sink. Type: string (or Expression with resultType string). | |
enableAdlsSingleFileParallel | Single File Parallel. |
AzureMySqlSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "AzureMySqlSink" (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). |
AzurePostgreSqlSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "AzurePostgreSqlSink" (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). |
AzureQueueSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "AzureQueueSink" (required) |
AzureSearchIndexSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "AzureSearchIndexSink" (required) |
writeBehavior | Specify the write behavior when upserting documents into Azure Search Index. | "Merge" "Upload" |
AzureSqlSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "AzureSqlSink" (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | |
sqlWriterStoredProcedureName | SQL writer stored procedure name. Type: string (or Expression with resultType string). | |
sqlWriterTableType | SQL writer table type. Type: string (or Expression with resultType string). | |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | |
storedProcedureParameters | SQL stored procedure parameters. | |
storedProcedureTableTypeParameterName | The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). | |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | |
upsertSettings | SQL upsert settings. | SqlUpsertSettings |
writeBehavior | Write behavior when copying data into Azure SQL. Type: SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum) |
SqlUpsertSettings
Name | Description | Value |
---|---|---|
interimSchemaName | Schema name for interim table. Type: string (or Expression with resultType string). | |
keys | Key column names for unique row identification. Type: array of strings (or Expression with resultType array of strings). | |
useTempDB | Specifies whether to use temp db for upsert interim table. Type: boolean (or Expression with resultType boolean). |
AzureTableSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "AzureTableSink" (required) |
azureTableDefaultPartitionKeyValue | Azure Table default partition key value. Type: string (or Expression with resultType string). | |
azureTableInsertType | Azure Table insert type. Type: string (or Expression with resultType string). | |
azureTablePartitionKeyName | Azure Table partition key name. Type: string (or Expression with resultType string). | |
azureTableRowKeyName | Azure Table row key name. Type: string (or Expression with resultType string). |
BinarySink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "BinarySink" (required) |
storeSettings | Binary store settings. | StoreWriteSettings |
BlobSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "BlobSink" (required) |
blobWriterAddHeader | Blob writer add header. Type: boolean (or Expression with resultType boolean). | |
blobWriterDateTimeFormat | Blob writer date time format. Type: string (or Expression with resultType string). | |
blobWriterOverwriteFiles | Blob writer overwrite files. Type: boolean (or Expression with resultType boolean). | |
copyBehavior | The type of copy behavior for copy sink. | |
metadata | Specify the custom metadata to be added to sink data. Type: array of objects (or Expression with resultType array of objects). | MetadataItem[] |
CommonDataServiceForAppsSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "CommonDataServiceForAppsSink" (required) |
alternateKeyName | The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether to ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. | "Upsert" (required) |
CosmosDbMongoDbApiSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "CosmosDbMongoDbApiSink" (required) |
writeBehavior | Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). |
CosmosDbSqlApiSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "CosmosDbSqlApiSink" (required) |
writeBehavior | Describes how to write data to Azure Cosmos DB. Type: string (or Expression with resultType string). Allowed values: insert and upsert. |
DelimitedTextSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "DelimitedTextSink" (required) |
formatSettings | DelimitedText format settings. | DelimitedTextWriteSettings |
storeSettings | DelimitedText store settings. | StoreWriteSettings |
DelimitedTextWriteSettings
Name | Description | Value |
---|---|---|
fileExtension | The file extension used to create the files. Type: string (or Expression with resultType string). | |
fileNamePrefix | Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). | |
maxRowsPerFile | Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). | |
quoteAllText | Indicates whether string values should always be enclosed with quotes. Type: boolean (or Expression with resultType boolean). | |
type | The write setting type. | string (required) |
{customized property} |
DocumentDbCollectionSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "DocumentDbCollectionSink" (required) |
nestingSeparator | Nested properties separator. Default is . (dot). Type: string (or Expression with resultType string). | |
writeBehavior | Describes how to write data to Azure Cosmos DB. Type: string (or Expression with resultType string). Allowed values: insert and upsert. |
DynamicsCrmSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "DynamicsCrmSink" (required) |
alternateKeyName | The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether to ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. | "Upsert" (required) |
DynamicsSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "DynamicsSink" (required) |
alternateKeyName | The logical name of the alternate key which will be used when upserting records. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. | "Upsert" (required) |
FileSystemSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "FileSystemSink" (required) |
copyBehavior | The type of copy behavior for copy sink. |
InformixSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "InformixSink" (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). |
JsonSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "JsonSink" (required) |
formatSettings | Json format settings. | JsonWriteSettings |
storeSettings | Json store settings. | StoreWriteSettings |
JsonWriteSettings
Name | Description | Value |
---|---|---|
filePattern | File pattern of JSON. This setting controls the way a collection of JSON objects will be treated. The default value is 'setOfObjects'. It is case-sensitive. | |
type | The write setting type. | string (required) |
{customized property} |
LakeHouseTableSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "LakeHouseTableSink" (required) |
partitionNameList | Specify the partition column names from sink columns. Type: array of objects (or Expression with resultType array of objects). | |
partitionOption | Create partitions in folder structure based on one or multiple columns. Each distinct column value (pair) will be a new partition. Possible values include: "None", "PartitionByKey". | |
tableActionOption | The type of table action for LakeHouse Table sink. Possible values include: "None", "Append", "Overwrite". |
MicrosoftAccessSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "MicrosoftAccessSink" (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). |
MongoDbAtlasSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "MongoDbAtlasSink" (required) |
writeBehavior | Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). |
MongoDbV2Sink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "MongoDbV2Sink" (required) |
writeBehavior | Specifies whether the document with same key to be overwritten (upsert) rather than throw exception (insert). The default value is "insert". Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). |
OdbcSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "OdbcSink" (required) |
preCopyScript | A query to execute before starting the copy. Type: string (or Expression with resultType string). |
OracleSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "OracleSink" (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). |
OrcSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "OrcSink" (required) |
formatSettings | ORC format settings. | OrcWriteSettings |
storeSettings | ORC store settings. | StoreWriteSettings |
OrcWriteSettings
Name | Description | Value |
---|---|---|
fileNamePrefix | Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). | |
maxRowsPerFile | Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). | |
type | The write setting type. | string (required) |
{customized property} |
ParquetSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "ParquetSink" (required) |
formatSettings | Parquet format settings. | ParquetWriteSettings |
storeSettings | Parquet store settings. | StoreWriteSettings |
ParquetWriteSettings
Name | Description | Value |
---|---|---|
fileNamePrefix | Specifies the file name pattern {fileNamePrefix}_{fileIndex}.{fileExtension} when copy from non-file based store without partitionOptions. Type: string (or Expression with resultType string). | |
maxRowsPerFile | Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). | |
type | The write setting type. | string (required) |
{customized property} |
RestSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "RestSink" (required) |
additionalHeaders | The additional HTTP headers in the request to the RESTful API. Type: key value pairs (value should be string type). | |
httpCompressionType | Http Compression Type to Send data in compressed format with Optimal Compression Level, Default is None. And The Only Supported option is Gzip. Type: string (or Expression with resultType string). | |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
requestInterval | The time to await before sending next request, in milliseconds | |
requestMethod | The HTTP method used to call the RESTful API. The default is POST. Type: string (or Expression with resultType string). |
SalesforceServiceCloudSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "SalesforceServiceCloudSink" (required) |
externalIdFieldName | The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. Default is Insert. | "Insert" "Upsert" |
SalesforceServiceCloudV2Sink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "SalesforceServiceCloudV2Sink" (required) |
externalIdFieldName | The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. Default is Insert. | "Insert" "Upsert" |
SalesforceSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "SalesforceSink" (required) |
externalIdFieldName | The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. Default is Insert. | "Insert" "Upsert" |
SalesforceV2Sink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "SalesforceV2Sink" (required) |
externalIdFieldName | The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). | |
ignoreNullValues | The flag indicating whether or not to ignore null values from input dataset (except key fields) during write operation. Default value is false. If set it to true, it means ADF will leave the data in the destination object unchanged when doing upsert/update operation and insert defined default value when doing insert operation, versus ADF will update the data in the destination object to NULL when doing upsert/update operation and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType boolean). | |
writeBehavior | The write behavior for the operation. Default is Insert. | "Insert" "Upsert" |
SapCloudForCustomerSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "SapCloudForCustomerSink" (required) |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
writeBehavior | The write behavior for the operation. Default is 'Insert'. | "Insert" "Update" |
SnowflakeSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "SnowflakeSink" (required) |
importSettings | Snowflake import settings. | SnowflakeImportCopyCommand |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). |
SnowflakeImportCopyCommand
Name | Description | Value |
---|---|---|
additionalCopyOptions | Additional copy options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": "'HH24:MI:SS.FF'" } | SnowflakeImportCopyCommandAdditionalCopyOptions |
additionalFormatOptions | Additional format options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalFormatOptions": { "FORCE": "TRUE", "LOAD_UNCERTAIN_FILES": "'FALSE'" } | SnowflakeImportCopyCommandAdditionalFormatOptions |
type | The import setting type. | string (required) |
{customized property} |
SnowflakeImportCopyCommandAdditionalCopyOptions
Name | Description | Value |
---|---|---|
{customized property} |
SnowflakeImportCopyCommandAdditionalFormatOptions
Name | Description | Value |
---|---|---|
{customized property} |
SnowflakeV2Sink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "SnowflakeV2Sink" (required) |
importSettings | Snowflake import settings. | SnowflakeImportCopyCommand |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). |
SqlDWSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "SqlDWSink" (required) |
allowCopyCommand | Indicates to use Copy Command to copy data into SQL Data Warehouse. Type: boolean (or Expression with resultType boolean). | |
allowPolyBase | Indicates to use PolyBase to copy data into SQL Data Warehouse when applicable. Type: boolean (or Expression with resultType boolean). | |
copyCommandSettings | Specifies Copy Command related settings when allowCopyCommand is true. | DWCopyCommandSettings |
polyBaseSettings | Specifies PolyBase-related settings when allowPolyBase is true. | PolybaseSettings |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | |
upsertSettings | SQL DW upsert settings. | SqlDWUpsertSettings |
writeBehavior | Write behavior when copying data into azure SQL DW. Type: SqlDWWriteBehaviorEnum (or Expression with resultType SqlDWWriteBehaviorEnum) |
DWCopyCommandSettings
Name | Description | Value |
---|---|---|
additionalOptions | Additional options directly passed to SQL DW in Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalOptions": { "MAXERRORS": "1000", "DATEFORMAT": "'ymd'" } | DWCopyCommandSettingsAdditionalOptions |
defaultValues | Specifies the default values for each target column in SQL DW. The default values in the property overwrite the DEFAULT constraint set in the DB, and identity column cannot have a default value. Type: array of objects (or Expression with resultType array of objects). | DWCopyCommandDefaultValue[] |
DWCopyCommandSettingsAdditionalOptions
Name | Description | Value |
---|---|---|
{customized property} | string |
DWCopyCommandDefaultValue
Name | Description | Value |
---|---|---|
columnName | Column name. Type: object (or Expression with resultType string). | |
defaultValue | The default value of the column. Type: object (or Expression with resultType string). |
PolybaseSettings
Name | Description | Value |
---|---|---|
rejectSampleValue | Determines the number of rows to attempt to retrieve before the PolyBase recalculates the percentage of rejected rows. Type: integer (or Expression with resultType integer), minimum: 0. | |
rejectType | Reject type. | "percentage" "value" |
rejectValue | Specifies the value or the percentage of rows that can be rejected before the query fails. Type: number (or Expression with resultType number), minimum: 0. | |
useTypeDefault | Specifies how to handle missing values in delimited text files when PolyBase retrieves data from the text file. Type: boolean (or Expression with resultType boolean). | |
{customized property} |
SqlDWUpsertSettings
Name | Description | Value |
---|---|---|
interimSchemaName | Schema name for interim table. Type: string (or Expression with resultType string). | |
keys | Key column names for unique row identification. Type: array of strings (or Expression with resultType array of strings). |
SqlMISink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "SqlMISink" (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | |
sqlWriterStoredProcedureName | SQL writer stored procedure name. Type: string (or Expression with resultType string). | |
sqlWriterTableType | SQL writer table type. Type: string (or Expression with resultType string). | |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | |
storedProcedureParameters | SQL stored procedure parameters. | |
storedProcedureTableTypeParameterName | The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). | |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | |
upsertSettings | SQL upsert settings. | SqlUpsertSettings |
writeBehavior | White behavior when copying data into azure SQL MI. Type: string (or Expression with resultType string) |
SqlServerSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "SqlServerSink" (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | |
sqlWriterStoredProcedureName | SQL writer stored procedure name. Type: string (or Expression with resultType string). | |
sqlWriterTableType | SQL writer table type. Type: string (or Expression with resultType string). | |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | |
storedProcedureParameters | SQL stored procedure parameters. | |
storedProcedureTableTypeParameterName | The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). | |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | |
upsertSettings | SQL upsert settings. | SqlUpsertSettings |
writeBehavior | Write behavior when copying data into sql server. Type: string (or Expression with resultType string). |
SqlSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "SqlSink" (required) |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | |
sqlWriterStoredProcedureName | SQL writer stored procedure name. Type: string (or Expression with resultType string). | |
sqlWriterTableType | SQL writer table type. Type: string (or Expression with resultType string). | |
sqlWriterUseTableLock | Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). | |
storedProcedureParameters | SQL stored procedure parameters. | |
storedProcedureTableTypeParameterName | The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). | |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | |
upsertSettings | SQL upsert settings. | SqlUpsertSettings |
writeBehavior | Write behavior when copying data into sql. Type: string (or Expression with resultType string). |
WarehouseSink
Name | Description | Value |
---|---|---|
type | Copy sink type. | "WarehouseSink" (required) |
allowCopyCommand | Indicates to use Copy Command to copy data into SQL Data Warehouse. Type: boolean (or Expression with resultType boolean). | |
copyCommandSettings | Specifies Copy Command related settings when allowCopyCommand is true. | DWCopyCommandSettings |
preCopyScript | SQL pre-copy script. Type: string (or Expression with resultType string). | |
tableOption | The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). | |
writeBehavior | Write behavior when copying data into azure Microsoft Fabric Data Warehouse. Type: DWWriteBehaviorEnum (or Expression with resultType DWWriteBehaviorEnum) |
SkipErrorFile
Name | Description | Value |
---|---|---|
dataInconsistency | Skip if source/sink file changed by other concurrent write. Default is false. Type: boolean (or Expression with resultType boolean). | |
fileMissing | Skip if file is deleted by other client during copy. Default is true. Type: boolean (or Expression with resultType boolean). |
CopySource
AmazonMWSSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "AmazonMWSSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
AmazonRdsForOracleSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "AmazonRdsForOracleSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
oracleReaderQuery | AmazonRdsForOracle reader query. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for AmazonRdsForOracle read in parallel. Type: string (or Expression with resultType string). | |
partitionSettings | The settings that will be leveraged for AmazonRdsForOracle source partitioning. | AmazonRdsForOraclePartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
AmazonRdsForOraclePartitionSettings
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionNames | Names of the physical partitions of AmazonRdsForOracle table. | |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). |
AmazonRdsForSqlServerSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "AmazonRdsForSqlServerSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
produceAdditionalTypes | Which additional types to produce. | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". |
SqlPartitionSettings
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column in integer or datetime type that will be used for proceeding partitioning. If not specified, the primary key of the table is auto-detected and used as the partition column. Type: string (or Expression with resultType string). | |
partitionLowerBound | The minimum value of the partition column for partition range splitting. This value is used to decide the partition stride, not for filtering the rows in table. All rows in the table or query result will be partitioned and copied. Type: string (or Expression with resultType string). | |
partitionUpperBound | The maximum value of the partition column for partition range splitting. This value is used to decide the partition stride, not for filtering the rows in table. All rows in the table or query result will be partitioned and copied. Type: string (or Expression with resultType string). |
AmazonRedshiftSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "AmazonRedshiftSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
redshiftUnloadSettings | The Amazon S3 settings needed for the interim Amazon S3 when copying from Amazon Redshift with unload. With this, data from Amazon Redshift source will be unloaded into S3 first and then copied into the targeted sink from the interim S3. | RedshiftUnloadSettings |
RedshiftUnloadSettings
Name | Description | Value |
---|---|---|
bucketName | The bucket of the interim Amazon S3 which will be used to store the unloaded data from Amazon Redshift source. The bucket must be in the same region as the Amazon Redshift source. Type: string (or Expression with resultType string). | |
s3LinkedServiceName | The name of the Amazon S3 linked service which will be used for the unload operation when copying from the Amazon Redshift source. | LinkedServiceReference (required) |
AvroSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "AvroSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
storeSettings | Avro store settings. | StoreReadSettings |
StoreReadSettings
Name | Description | Value |
---|---|---|
disableMetricsCollection | If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). | |
maxConcurrentConnections | The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). | |
type | Set the object type | AmazonS3CompatibleReadSettings AmazonS3ReadSettings AzureBlobFSReadSettings AzureBlobStorageReadSettings AzureDataLakeStoreReadSettings AzureFileStorageReadSettings FileServerReadSettings FtpReadSettings GoogleCloudStorageReadSettings HdfsReadSettings HttpReadSettings LakeHouseReadSettings OracleCloudStorageReadSettings SftpReadSettings (required) |
AmazonS3CompatibleReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | "AmazonS3CompatibleReadSettings" (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
prefix | The prefix filter for the S3 Compatible object name. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Amazon S3 Compatible wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Amazon S3 Compatible wildcardFolderPath. Type: string (or Expression with resultType string). |
AmazonS3ReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | "AmazonS3ReadSettings" (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
prefix | The prefix filter for the S3 object name. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | AmazonS3 wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | AmazonS3 wildcardFolderPath. Type: string (or Expression with resultType string). |
AzureBlobFSReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | "AzureBlobFSReadSettings" (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Azure blobFS wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Azure blobFS wildcardFolderPath. Type: string (or Expression with resultType string). |
AzureBlobStorageReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | "AzureBlobStorageReadSettings" (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
prefix | The prefix filter for the Azure Blob name. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Azure blob wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Azure blob wildcardFolderPath. Type: string (or Expression with resultType string). |
AzureDataLakeStoreReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | "AzureDataLakeStoreReadSettings" (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
listAfter | Lists files after the value (exclusive) based on file/folder names’ lexicographical order. Applies under the folderPath in data set, and filter files/sub-folders under the folderPath. Type: string (or Expression with resultType string). | |
listBefore | Lists files before the value (inclusive) based on file/folder names’ lexicographical order. Applies under the folderPath in data set, and filter files/sub-folders under the folderPath. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | ADLS wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | ADLS wildcardFolderPath. Type: string (or Expression with resultType string). |
AzureFileStorageReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | "AzureFileStorageReadSettings" (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
prefix | The prefix filter for the Azure File name starting from root path. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Azure File Storage wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Azure File Storage wildcardFolderPath. Type: string (or Expression with resultType string). |
FileServerReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | "FileServerReadSettings" (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileFilter | Specify a filter to be used to select a subset of files in the folderPath rather than all files. Type: string (or Expression with resultType string). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | FileServer wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | FileServer wildcardFolderPath. Type: string (or Expression with resultType string). |
FtpReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | "FtpReadSettings" (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
disableChunking | If true, disable parallel reading within each file. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
useBinaryTransfer | Specify whether to use binary transfer mode for FTP stores. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Ftp wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Ftp wildcardFolderPath. Type: string (or Expression with resultType string). |
GoogleCloudStorageReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | "GoogleCloudStorageReadSettings" (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
prefix | The prefix filter for the Google Cloud Storage object name. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Google Cloud Storage wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Google Cloud Storage wildcardFolderPath. Type: string (or Expression with resultType string). |
HdfsReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | "HdfsReadSettings" (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
distcpSettings | Specifies Distcp-related settings. | DistcpSettings |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | HDFS wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | HDFS wildcardFolderPath. Type: string (or Expression with resultType string). |
DistcpSettings
Name | Description | Value |
---|---|---|
distcpOptions | Specifies the Distcp options. Type: string (or Expression with resultType string). | |
resourceManagerEndpoint | Specifies the Yarn ResourceManager endpoint. Type: string (or Expression with resultType string). | |
tempScriptPath | Specifies an existing folder path which will be used to store temp Distcp command script. The script file is generated by ADF and will be removed after Copy job finished. Type: string (or Expression with resultType string). |
HttpReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | "HttpReadSettings" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
additionalHeaders | The additional HTTP headers in the request to the RESTful API. Type: string (or Expression with resultType string). | |
requestBody | The HTTP request body to the RESTful API if requestMethod is POST. Type: string (or Expression with resultType string). | |
requestMethod | The HTTP method used to call the RESTful API. The default is GET. Type: string (or Expression with resultType string). | |
requestTimeout | Specifies the timeout for a HTTP client to get HTTP response from HTTP server. Type: string (or Expression with resultType string). |
LakeHouseReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | "LakeHouseReadSettings" (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Microsoft Fabric LakeHouse Files wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Microsoft Fabric LakeHouse Files wildcardFolderPath. Type: string (or Expression with resultType string). |
OracleCloudStorageReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | "OracleCloudStorageReadSettings" (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
prefix | The prefix filter for the Oracle Cloud Storage object name. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Oracle Cloud Storage wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Oracle Cloud Storage wildcardFolderPath. Type: string (or Expression with resultType string). |
SftpReadSettings
Name | Description | Value |
---|---|---|
type | The read setting type. | "SftpReadSettings" (required) |
deleteFilesAfterCompletion | Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). | |
disableChunking | If true, disable parallel reading within each file. Default is false. Type: boolean (or Expression with resultType boolean). | |
enablePartitionDiscovery | Indicates whether to enable partition discovery. Type: boolean (or Expression with resultType boolean). | |
fileListPath | Point to a text file that lists each file (relative path to the path configured in the dataset) that you want to copy. Type: string (or Expression with resultType string). | |
modifiedDatetimeEnd | The end of file's modified datetime. Type: string (or Expression with resultType string). | |
modifiedDatetimeStart | The start of file's modified datetime. Type: string (or Expression with resultType string). | |
partitionRootPath | Specify the root path where partition discovery starts from. Type: string (or Expression with resultType string). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
wildcardFileName | Sftp wildcardFileName. Type: string (or Expression with resultType string). | |
wildcardFolderPath | Sftp wildcardFolderPath. Type: string (or Expression with resultType string). |
AzureBlobFSSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "AzureBlobFSSource" (required) |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
skipHeaderLineCount | Number of header lines to skip from each blob. Type: integer (or Expression with resultType integer). | |
treatEmptyAsNull | Treat empty as null. Type: boolean (or Expression with resultType boolean). |
AzureDatabricksDeltaLakeSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "AzureDatabricksDeltaLakeSource" (required) |
exportSettings | Azure Databricks Delta Lake export settings. | AzureDatabricksDeltaLakeExportCommand |
query | Azure Databricks Delta Lake Sql query. Type: string (or Expression with resultType string). |
AzureDatabricksDeltaLakeExportCommand
Name | Description | Value |
---|---|---|
dateFormat | Specify the date format for the csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). | |
timestampFormat | Specify the timestamp format for the csv in Azure Databricks Delta Lake Copy. Type: string (or Expression with resultType string). | |
type | The export setting type. | string (required) |
{customized property} |
AzureDataExplorerSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "AzureDataExplorerSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
noTruncation | The name of the Boolean option that controls whether truncation is applied to result-sets that go beyond a certain row-count limit. | |
query | Database query. Should be a Kusto Query Language (KQL) query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])).. |
AzureDataLakeStoreSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "AzureDataLakeStoreSource" (required) |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). |
AzureMariaDBSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "AzureMariaDBSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
AzureMySqlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "AzureMySqlSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
AzurePostgreSqlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "AzurePostgreSqlSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
AzureSqlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "AzureSqlSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
produceAdditionalTypes | Which additional types to produce. | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". |
AzureTableSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "AzureTableSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
azureTableSourceIgnoreTableNotFound | Azure Table source ignore table not found. Type: boolean (or Expression with resultType boolean). | |
azureTableSourceQuery | Azure Table source query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
BinarySource
Name | Description | Value |
---|---|---|
type | Copy source type. | "BinarySource" (required) |
formatSettings | Binary format settings. | BinaryReadSettings |
storeSettings | Binary store settings. | StoreReadSettings |
BinaryReadSettings
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
type | The read setting type. | string (required) |
{customized property} |
CompressionReadSettings
Name | Description | Value |
---|---|---|
type | Set the object type | TarGZipReadSettings TarReadSettings ZipDeflateReadSettings (required) |
TarGZipReadSettings
Name | Description | Value |
---|---|---|
type | The Compression setting type. | "TarGZipReadSettings" (required) |
preserveCompressionFileNameAsFolder | Preserve the compression file name as folder path. Type: boolean (or Expression with resultType boolean). |
TarReadSettings
Name | Description | Value |
---|---|---|
type | The Compression setting type. | "TarReadSettings" (required) |
preserveCompressionFileNameAsFolder | Preserve the compression file name as folder path. Type: boolean (or Expression with resultType boolean). |
ZipDeflateReadSettings
Name | Description | Value |
---|---|---|
type | The Compression setting type. | "ZipDeflateReadSettings" (required) |
preserveZipFileNameAsFolder | Preserve the zip file name as folder path. Type: boolean (or Expression with resultType boolean). |
BlobSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "BlobSource" (required) |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). | |
skipHeaderLineCount | Number of header lines to skip from each blob. Type: integer (or Expression with resultType integer). | |
treatEmptyAsNull | Treat empty as null. Type: boolean (or Expression with resultType boolean). |
CassandraSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "CassandraSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
consistencyLevel | The consistency level specifies how many Cassandra servers must respond to a read request before returning data to the client application. Cassandra checks the specified number of Cassandra servers for data to satisfy the read request. Must be one of cassandraSourceReadConsistencyLevels. The default value is 'ONE'. It is case-insensitive. | "ALL" "EACH_QUORUM" "LOCAL_ONE" "LOCAL_QUORUM" "LOCAL_SERIAL" "ONE" "QUORUM" "SERIAL" "THREE" "TWO" |
query | Database query. Should be a SQL-92 query expression or Cassandra Query Language (CQL) command. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
CommonDataServiceForAppsSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "CommonDataServiceForAppsSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | FetchXML is a proprietary query language that is used in Microsoft Common Data Service for Apps (online & on-premises). Type: string (or Expression with resultType string). |
ConcurSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "ConcurSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
CosmosDbMongoDbApiSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "CosmosDbMongoDbApiSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
batchSize | Specifies the number of documents to return in each batch of the response from MongoDB instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer). | |
cursorMethods | Cursor methods for Mongodb query. | MongoDbCursorMethodsProperties |
filter | Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
MongoDbCursorMethodsProperties
Name | Description | Value |
---|---|---|
limit | Specifies the maximum number of documents the server returns. limit() is analogous to the LIMIT statement in a SQL database. Type: integer (or Expression with resultType integer). | |
project | Specifies the fields to return in the documents that match the query filter. To return all fields in the matching documents, omit this parameter. Type: string (or Expression with resultType string). | |
skip | Specifies the how many documents skipped and where MongoDB begins returning results. This approach may be useful in implementing paginated results. Type: integer (or Expression with resultType integer). | |
sort | Specifies the order in which the query returns matching documents. Type: string (or Expression with resultType string). Type: string (or Expression with resultType string). | |
{customized property} |
CosmosDbSqlApiSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "CosmosDbSqlApiSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
detectDatetime | Whether detect primitive values as datetime values. Type: boolean (or Expression with resultType boolean). | |
pageSize | Page size of the result. Type: integer (or Expression with resultType integer). | |
preferredRegions | Preferred regions. Type: array of strings (or Expression with resultType array of strings). | |
query | SQL API query. Type: string (or Expression with resultType string). |
CouchbaseSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "CouchbaseSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Db2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | "Db2Source" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
DelimitedTextSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "DelimitedTextSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
formatSettings | DelimitedText format settings. | DelimitedTextReadSettings |
storeSettings | DelimitedText store settings. | StoreReadSettings |
DelimitedTextReadSettings
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
skipLineCount | Indicates the number of non-empty rows to skip when reading data from input files. Type: integer (or Expression with resultType integer). | |
type | The read setting type. | string (required) |
{customized property} |
DocumentDbCollectionSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "DocumentDbCollectionSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
nestingSeparator | Nested properties separator. Type: string (or Expression with resultType string). | |
query | Documents query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
DrillSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "DrillSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
DynamicsAXSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "DynamicsAXSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
DynamicsCrmSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "DynamicsCrmSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | FetchXML is a proprietary query language that is used in Microsoft Dynamics CRM (online & on-premises). Type: string (or Expression with resultType string). |
DynamicsSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "DynamicsSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | FetchXML is a proprietary query language that is used in Microsoft Dynamics (online & on-premises). Type: string (or Expression with resultType string). |
EloquaSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "EloquaSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
ExcelSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "ExcelSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
storeSettings | Excel store settings. | StoreReadSettings |
FileSystemSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "FileSystemSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). |
GoogleAdWordsSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "GoogleAdWordsSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
GoogleBigQuerySource
Name | Description | Value |
---|---|---|
type | Copy source type. | "GoogleBigQuerySource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
GoogleBigQueryV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | "GoogleBigQueryV2Source" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
GreenplumSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "GreenplumSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
HBaseSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "HBaseSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
HdfsSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "HdfsSource" (required) |
distcpSettings | Specifies Distcp-related settings. | DistcpSettings |
recursive | If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). |
HiveSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "HiveSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
HttpSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "HttpSource" (required) |
httpRequestTimeout | Specifies the timeout for a HTTP client to get HTTP response from HTTP server. The default value is equivalent to System.Net.HttpWebRequest.Timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
HubspotSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "HubspotSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
ImpalaSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "ImpalaSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
InformixSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "InformixSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
JiraSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "JiraSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
JsonSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "JsonSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
formatSettings | Json format settings. | JsonReadSettings |
storeSettings | Json store settings. | StoreReadSettings |
JsonReadSettings
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
type | The read setting type. | string (required) |
{customized property} |
LakeHouseTableSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "LakeHouseTableSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
timestampAsOf | Query an older snapshot by timestamp. Type: string (or Expression with resultType string). | |
versionAsOf | Query an older snapshot by version. Type: integer (or Expression with resultType integer). |
MagentoSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "MagentoSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
MariaDBSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "MariaDBSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
MarketoSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "MarketoSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
MicrosoftAccessSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "MicrosoftAccessSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). |
MongoDbAtlasSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "MongoDbAtlasSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
batchSize | Specifies the number of documents to return in each batch of the response from MongoDB Atlas instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer). | |
cursorMethods | Cursor methods for Mongodb query | MongoDbCursorMethodsProperties |
filter | Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
MongoDbSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "MongoDbSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Should be a SQL-92 query expression. Type: string (or Expression with resultType string). |
MongoDbV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | "MongoDbV2Source" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
batchSize | Specifies the number of documents to return in each batch of the response from MongoDB instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. Type: integer (or Expression with resultType integer). | |
cursorMethods | Cursor methods for Mongodb query | MongoDbCursorMethodsProperties |
filter | Specifies selection filter using query operators. To return all documents in a collection, omit this parameter or pass an empty document ({}). Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
MySqlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "MySqlSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
NetezzaSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "NetezzaSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
partitionOption | The partition mechanism that will be used for Netezza read in parallel. Possible values include: "None", "DataSlice", "DynamicRange". | |
partitionSettings | The settings that will be leveraged for Netezza source partitioning. | NetezzaPartitionSettings |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
NetezzaPartitionSettings
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). |
ODataSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "ODataSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
query | OData query. For example, "$top=1". Type: string (or Expression with resultType string). |
OdbcSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "OdbcSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
Office365Source
Name | Description | Value |
---|---|---|
type | Copy source type. | "Office365Source" (required) |
allowedGroups | The groups containing all the users. Type: array of strings (or Expression with resultType array of strings). | |
dateFilterColumn | The Column to apply the {paramref name="StartTime"/} and {paramref name="EndTime"/}. Type: string (or Expression with resultType string). | |
endTime | End time of the requested range for this dataset. Type: string (or Expression with resultType string). | |
outputColumns | The columns to be read out from the Office 365 table. Type: array of objects (or Expression with resultType array of objects). itemType: OutputColumn. Example: [ { "name": "Id" }, { "name": "CreatedDateTime" } ] | |
startTime | Start time of the requested range for this dataset. Type: string (or Expression with resultType string). | |
userScopeFilterUri | The user scope uri. Type: string (or Expression with resultType string). |
OracleServiceCloudSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "OracleServiceCloudSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
OracleSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "OracleSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
oracleReaderQuery | Oracle reader query. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Oracle read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". | |
partitionSettings | The settings that will be leveraged for Oracle source partitioning. | OraclePartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
OraclePartitionSettings
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column in integer type that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionNames | Names of the physical partitions of Oracle table. | |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). |
OrcSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "OrcSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
storeSettings | ORC store settings. | StoreReadSettings |
ParquetSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "ParquetSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
formatSettings | Parquet format settings. | ParquetReadSettings |
storeSettings | Parquet store settings. | StoreReadSettings |
ParquetReadSettings
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
type | The read setting type. | string (required) |
{customized property} |
PaypalSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "PaypalSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
PhoenixSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "PhoenixSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
PostgreSqlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "PostgreSqlSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
PostgreSqlV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | "PostgreSqlV2Source" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
PrestoSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "PrestoSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
QuickBooksSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "QuickBooksSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
RelationalSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "RelationalSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). |
ResponsysSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "ResponsysSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
RestSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "RestSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: key value pairs (value should be string type). | |
additionalHeaders | The additional HTTP headers in the request to the RESTful API. Type: string (or Expression with resultType string). | |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
paginationRules | The pagination rules to compose next page requests. Type: string (or Expression with resultType string). | |
requestBody | The HTTP request body to the RESTful API if requestMethod is POST. Type: string (or Expression with resultType string). | |
requestInterval | The time to await before sending next page request. | |
requestMethod | The HTTP method used to call the RESTful API. The default is GET. Type: string (or Expression with resultType string). |
SalesforceMarketingCloudSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SalesforceMarketingCloudSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
SalesforceServiceCloudSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SalesforceServiceCloudSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
readBehavior | The read behavior for the operation. Default is Query. Allowed values: Query/QueryAll. Type: string (or Expression with resultType string). |
SalesforceServiceCloudV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | "SalesforceServiceCloudV2Source" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
includeDeletedObjects | This property control whether query result contains Deleted objects. Default is false. Type: boolean (or Expression with resultType boolean). | |
SOQLQuery | Database query. Type: string (or Expression with resultType string). |
SalesforceSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SalesforceSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
readBehavior | The read behavior for the operation. Default is Query. Allowed values: Query/QueryAll. Type: string (or Expression with resultType string). |
SalesforceV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | "SalesforceV2Source" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
includeDeletedObjects | This property control whether query result contains Deleted objects. Default is false. Type: boolean (or Expression with resultType boolean). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
SOQLQuery | Database query. Type: string (or Expression with resultType string). |
SapBwSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SapBwSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | MDX query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
SapCloudForCustomerSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SapCloudForCustomerSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
query | SAP Cloud for Customer OData query. For example, "$top=1". Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
SapEccSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SapEccSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
httpRequestTimeout | The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
query | SAP ECC OData query. For example, "$top=1". Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
SapHanaSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SapHanaSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
packetSize | The packet size of data read from SAP HANA. Type: integer(or Expression with resultType integer). | |
partitionOption | The partition mechanism that will be used for SAP HANA read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "SapHanaDynamicRange". | |
partitionSettings | The settings that will be leveraged for SAP HANA source partitioning. | SapHanaPartitionSettings |
query | SAP HANA Sql query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
SapHanaPartitionSettings
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). |
SapOdpSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SapOdpSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
extractionMode | The extraction mode. Allowed value include: Full, Delta and Recovery. The default value is Full. Type: string (or Expression with resultType string). | |
projection | Specifies the columns to be selected from source data. Type: array of objects(projection) (or Expression with resultType array of objects). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
selection | Specifies the selection conditions from source data. Type: array of objects(selection) (or Expression with resultType array of objects). | |
subscriberProcess | The subscriber process to manage the delta process. Type: string (or Expression with resultType string). |
SapOpenHubSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SapOpenHubSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
baseRequestId | The ID of request for delta loading. Once it is set, only data with requestId larger than the value of this property will be retrieved. The default value is 0. Type: integer (or Expression with resultType integer ). | |
customRfcReadTableFunctionModule | Specifies the custom RFC function module that will be used to read data from SAP Table. Type: string (or Expression with resultType string). | |
excludeLastRequest | Whether to exclude the records of the last request. The default value is true. Type: boolean (or Expression with resultType boolean). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sapDataColumnDelimiter | The single character that will be used as delimiter passed to SAP RFC as well as splitting the output data retrieved. Type: string (or Expression with resultType string). |
SapTableSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SapTableSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
batchSize | Specifies the maximum number of rows that will be retrieved at a time when retrieving data from SAP Table. Type: integer (or Expression with resultType integer). | |
customRfcReadTableFunctionModule | Specifies the custom RFC function module that will be used to read data from SAP Table. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for SAP table read in parallel. Possible values include: "None", "PartitionOnInt", "PartitionOnCalendarYear", "PartitionOnCalendarMonth", "PartitionOnCalendarDate", "PartitionOnTime". | |
partitionSettings | The settings that will be leveraged for SAP table source partitioning. | SapTablePartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
rfcTableFields | The fields of the SAP table that will be retrieved. For example, column0, column1. Type: string (or Expression with resultType string). | |
rfcTableOptions | The options for the filtering of the SAP Table. For example, COLUMN0 EQ SOME VALUE. Type: string (or Expression with resultType string). | |
rowCount | The number of rows to be retrieved. Type: integer(or Expression with resultType integer). | |
rowSkips | The number of rows that will be skipped. Type: integer (or Expression with resultType integer). | |
sapDataColumnDelimiter | The single character that will be used as delimiter passed to SAP RFC as well as splitting the output data retrieved. Type: string (or Expression with resultType string). |
SapTablePartitionSettings
Name | Description | Value |
---|---|---|
maxPartitionsNumber | The maximum value of partitions the table will be split into. Type: integer (or Expression with resultType string). | |
partitionColumnName | The name of the column that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). |
ServiceNowSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "ServiceNowSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
ServiceNowV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | "ServiceNowV2Source" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
expression | Expression to filter data from source. | ExpressionV2 |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
ExpressionV2
Name | Description | Value |
---|---|---|
operands | List of nested expressions. | ExpressionV2[] |
operator | Expression operator value Type: string. | string |
type | Type of expressions supported by the system. Type: string. | "Binary" "Constant" "Field" "Unary" |
value | Value for Constant/Field Type: string. | string |
SharePointOnlineListSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SharePointOnlineListSource" (required) |
httpRequestTimeout | The wait time to get a response from SharePoint Online. Default value is 5 minutes (00:05:00). Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
query | The OData query to filter the data in SharePoint Online list. For example, "$top=1". Type: string (or Expression with resultType string). |
ShopifySource
Name | Description | Value |
---|---|---|
type | Copy source type. | "ShopifySource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
SnowflakeSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SnowflakeSource" (required) |
exportSettings | Snowflake export settings. | SnowflakeExportCopyCommand (required) |
query | Snowflake Sql query. Type: string (or Expression with resultType string). |
SnowflakeExportCopyCommand
Name | Description | Value |
---|---|---|
additionalCopyOptions | Additional copy options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": "'HH24:MI:SS.FF'" } | SnowflakeExportCopyCommandAdditionalCopyOptions |
additionalFormatOptions | Additional format options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalFormatOptions": { "OVERWRITE": "TRUE", "MAX_FILE_SIZE": "'FALSE'" } | SnowflakeExportCopyCommandAdditionalFormatOptions |
type | The export setting type. | string (required) |
{customized property} |
SnowflakeExportCopyCommandAdditionalCopyOptions
Name | Description | Value |
---|---|---|
{customized property} |
SnowflakeExportCopyCommandAdditionalFormatOptions
Name | Description | Value |
---|---|---|
{customized property} |
SnowflakeV2Source
Name | Description | Value |
---|---|---|
type | Copy source type. | "SnowflakeV2Source" (required) |
exportSettings | Snowflake export settings. | SnowflakeExportCopyCommand (required) |
query | Snowflake Sql query. Type: string (or Expression with resultType string). |
SparkSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SparkSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
SqlDWSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SqlDWSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | SQL Data Warehouse reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Data Warehouse source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". Type: object (or Expression with resultType object), itemType: StoredProcedureParameter. |
SqlMISource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SqlMISource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
produceAdditionalTypes | Which additional types to produce. | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a Azure SQL Managed Instance source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". |
SqlServerSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SqlServerSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
produceAdditionalTypes | Which additional types to produce. | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". |
SqlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SqlSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or Expression with resultType string). | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | SQL reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a SQL Database source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". |
SquareSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SquareSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
SybaseSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "SybaseSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | Database query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
TeradataSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "TeradataSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
partitionOption | The partition mechanism that will be used for teradata read in parallel. Possible values include: "None", "Hash", "DynamicRange". | |
partitionSettings | The settings that will be leveraged for teradata source partitioning. | TeradataPartitionSettings |
query | Teradata query. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
TeradataPartitionSettings
Name | Description | Value |
---|---|---|
partitionColumnName | The name of the column that will be used for proceeding range or hash partitioning. Type: string (or Expression with resultType string). | |
partitionLowerBound | The minimum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). | |
partitionUpperBound | The maximum value of column specified in partitionColumnName that will be used for proceeding range partitioning. Type: string (or Expression with resultType string). |
VerticaSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "VerticaSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
WarehouseSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "WarehouseSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
isolationLevel | Specifies the transaction locking behavior for the Microsoft Fabric Warehouse source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). | |
partitionOption | The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". | |
partitionSettings | The settings that will be leveraged for Sql source partitioning. | SqlPartitionSettings |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
sqlReaderQuery | Microsoft Fabric Warehouse reader query. Type: string (or Expression with resultType string). | |
sqlReaderStoredProcedureName | Name of the stored procedure for a Microsoft Fabric Warehouse source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". Type: object (or Expression with resultType object), itemType: StoredProcedureParameter. |
WebSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "WebSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). |
XeroSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "XeroSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
XmlSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "XmlSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
formatSettings | Xml format settings. | XmlReadSettings |
storeSettings | Xml store settings. | StoreReadSettings |
XmlReadSettings
Name | Description | Value |
---|---|---|
compressionProperties | Compression settings. | CompressionReadSettings |
detectDataType | Indicates whether type detection is enabled when reading the xml files. Type: boolean (or Expression with resultType boolean). | |
namespacePrefixes | Namespace uri to prefix mappings to override the prefixes in column names when namespace is enabled, if no prefix is defined for a namespace uri, the prefix of xml element/attribute name in the xml data file will be used. Example: "{"http://www.example.com/xml":"prefix"}" Type: object (or Expression with resultType object). | |
namespaces | Indicates whether namespace is enabled when reading the xml files. Type: boolean (or Expression with resultType boolean). | |
type | The read setting type. | string (required) |
validationMode | Indicates what validation method is used when reading the xml files. Allowed values: 'none', 'xsd', or 'dtd'. Type: string (or Expression with resultType string). | |
{customized property} |
ZohoSource
Name | Description | Value |
---|---|---|
type | Copy source type. | "ZohoSource" (required) |
additionalColumns | Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). | |
query | A query to retrieve data from source. Type: string (or Expression with resultType string). | |
queryTimeout | Query timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
StagingSettings
Name | Description | Value |
---|---|---|
enableCompression | Specifies whether to use compression when copying data via an interim staging. Default value is false. Type: boolean (or Expression with resultType boolean). | |
linkedServiceName | Staging linked service reference. | LinkedServiceReference (required) |
path | The path to storage for storing the interim data. Type: string (or Expression with resultType string). | |
{customized property} |
CustomActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "Custom" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Custom activity properties. | CustomActivityTypeProperties (required) |
CustomActivityTypeProperties
Name | Description | Value |
---|---|---|
autoUserSpecification | Elevation level and scope for the user, default is nonadmin task. Type: string (or Expression with resultType double). | |
command | Command for custom activity Type: string (or Expression with resultType string). | |
extendedProperties | User defined property bag. There is no restriction on the keys or values that can be used. The user specified custom activity has the full responsibility to consume and interpret the content defined. | CustomActivityTypePropertiesExtendedProperties |
folderPath | Folder path for resource files Type: string (or Expression with resultType string). | |
referenceObjects | Reference objects | CustomActivityReferenceObject |
resourceLinkedService | Resource linked service reference. | LinkedServiceReference |
retentionTimeInDays | The retention time for the files submitted for custom activity. Type: double (or Expression with resultType double). |
CustomActivityTypePropertiesExtendedProperties
Name | Description | Value |
---|---|---|
{customized property} |
CustomActivityReferenceObject
Name | Description | Value |
---|---|---|
datasets | Dataset references. | DatasetReference[] |
linkedServices | Linked service references. | LinkedServiceReference[] |
DatabricksNotebookActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "DatabricksNotebook" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Databricks Notebook activity properties. | DatabricksNotebookActivityTypeProperties (required) |
DatabricksNotebookActivityTypeProperties
Name | Description | Value |
---|---|---|
baseParameters | Base parameters to be used for each run of this job.If the notebook takes a parameter that is not specified, the default value from the notebook will be used. | DatabricksNotebookActivityTypePropertiesBaseParamete... |
libraries | A list of libraries to be installed on the cluster that will execute the job. | DatabricksNotebookActivityTypePropertiesLibrariesIte...[] |
notebookPath | The absolute path of the notebook to be run in the Databricks Workspace. This path must begin with a slash. Type: string (or Expression with resultType string). |
DatabricksNotebookActivityTypePropertiesBaseParamete...
Name | Description | Value |
---|---|---|
{customized property} |
DatabricksNotebookActivityTypePropertiesLibrariesIte...
Name | Description | Value |
---|---|---|
{customized property} |
DatabricksSparkJarActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "DatabricksSparkJar" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Databricks SparkJar activity properties. | DatabricksSparkJarActivityTypeProperties (required) |
DatabricksSparkJarActivityTypeProperties
Name | Description | Value |
---|---|---|
libraries | A list of libraries to be installed on the cluster that will execute the job. | DatabricksSparkJarActivityTypePropertiesLibrariesIte...[] |
mainClassName | The full name of the class containing the main method to be executed. This class must be contained in a JAR provided as a library. Type: string (or Expression with resultType string). | |
parameters | Parameters that will be passed to the main method. | any[] |
DatabricksSparkJarActivityTypePropertiesLibrariesIte...
Name | Description | Value |
---|---|---|
{customized property} |
DatabricksSparkPythonActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "DatabricksSparkPython" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Databricks SparkPython activity properties. | DatabricksSparkPythonActivityTypeProperties (required) |
DatabricksSparkPythonActivityTypeProperties
Name | Description | Value |
---|---|---|
libraries | A list of libraries to be installed on the cluster that will execute the job. | DatabricksSparkPythonActivityTypePropertiesLibraries...[] |
parameters | Command line parameters that will be passed to the Python file. | any[] |
pythonFile | The URI of the Python file to be executed. DBFS paths are supported. Type: string (or Expression with resultType string). |
DatabricksSparkPythonActivityTypePropertiesLibraries...
Name | Description | Value |
---|---|---|
{customized property} |
DataLakeAnalyticsUsqlActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "DataLakeAnalyticsU-SQL" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Data Lake Analytics U-SQL activity properties. | DataLakeAnalyticsUsqlActivityTypeProperties (required) |
DataLakeAnalyticsUsqlActivityTypeProperties
Name | Description | Value |
---|---|---|
compilationMode | Compilation mode of U-SQL. Must be one of these values : Semantic, Full and SingleBox. Type: string (or Expression with resultType string). | |
degreeOfParallelism | The maximum number of nodes simultaneously used to run the job. Default value is 1. Type: integer (or Expression with resultType integer), minimum: 1. | |
parameters | Parameters for U-SQL job request. | DataLakeAnalyticsUsqlActivityTypePropertiesParameter... |
priority | Determines which jobs out of all that are queued should be selected to run first. The lower the number, the higher the priority. Default value is 1000. Type: integer (or Expression with resultType integer), minimum: 1. | |
runtimeVersion | Runtime version of the U-SQL engine to use. Type: string (or Expression with resultType string). | |
scriptLinkedService | Script linked service reference. | LinkedServiceReference (required) |
scriptPath | Case-sensitive path to folder that contains the U-SQL script. Type: string (or Expression with resultType string). |
DataLakeAnalyticsUsqlActivityTypePropertiesParameter...
Name | Description | Value |
---|---|---|
{customized property} |
DeleteActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "Delete" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Delete activity properties. | DeleteActivityTypeProperties (required) |
DeleteActivityTypeProperties
Name | Description | Value |
---|---|---|
dataset | Delete activity dataset reference. | DatasetReference (required) |
enableLogging | Whether to record detailed logs of delete-activity execution. Default value is false. Type: boolean (or Expression with resultType boolean). | |
logStorageSettings | Log storage settings customer need to provide when enableLogging is true. | LogStorageSettings |
maxConcurrentConnections | The max concurrent connections to connect data source at the same time. | int Constraints: Min value = 1 |
recursive | If true, files or sub-folders under current folder path will be deleted recursively. Default is false. Type: boolean (or Expression with resultType boolean). | |
storeSettings | Delete activity store settings. | StoreReadSettings |
ExecuteDataFlowActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "ExecuteDataFlow" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute data flow activity properties. | ExecuteDataFlowActivityTypeProperties (required) |
ExecuteDataFlowActivityTypeProperties
Name | Description | Value |
---|---|---|
compute | Compute properties for data flow activity. | ExecuteDataFlowActivityTypePropertiesCompute |
continueOnError | Continue on error setting used for data flow execution. Enables processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean) | |
dataFlow | Data flow reference. | DataFlowReference (required) |
integrationRuntime | The integration runtime reference. | IntegrationRuntimeReference |
runConcurrently | Concurrent run setting used for data flow execution. Allows sinks with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean) | |
sourceStagingConcurrency | Specify number of parallel staging for sources applicable to the sink. Type: integer (or Expression with resultType integer) | |
staging | Staging info for execute data flow activity. | DataFlowStagingInfo |
traceLevel | Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string) |
ExecuteDataFlowActivityTypePropertiesCompute
Name | Description | Value |
---|---|---|
computeType | Compute type of the cluster which will execute data flow job. Possible values include: 'General', 'MemoryOptimized', 'ComputeOptimized'. Type: string (or Expression with resultType string) | |
coreCount | Core count of the cluster which will execute data flow job. Supported values are: 8, 16, 32, 48, 80, 144 and 272. Type: integer (or Expression with resultType integer) |
DataFlowReference
Name | Description | Value |
---|---|---|
datasetParameters | Reference data flow parameters from dataset. | |
parameters | Data flow parameters | ParameterValueSpecification |
referenceName | Reference data flow name. | string (required) |
type | Data flow reference type. | "DataFlowReference" (required) |
{customized property} |
IntegrationRuntimeReference
Name | Description | Value |
---|---|---|
parameters | Arguments for integration runtime. | ParameterValueSpecification |
referenceName | Reference integration runtime name. | string (required) |
type | Type of integration runtime. | "IntegrationRuntimeReference" (required) |
DataFlowStagingInfo
Name | Description | Value |
---|---|---|
folderPath | Folder path for staging blob. Type: string (or Expression with resultType string) | |
linkedService | Staging linked service reference. | LinkedServiceReference |
ExecutePipelineActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "ExecutePipeline" (required) |
policy | Execute pipeline activity policy. | ExecutePipelineActivityPolicy |
typeProperties | Execute pipeline activity properties. | ExecutePipelineActivityTypeProperties (required) |
ExecutePipelineActivityPolicy
Name | Description | Value |
---|---|---|
secureInput | When set to true, Input from activity is considered as secure and will not be logged to monitoring. | bool |
{customized property} |
ExecutePipelineActivityTypeProperties
Name | Description | Value |
---|---|---|
parameters | Pipeline parameters. | ParameterValueSpecification |
pipeline | Pipeline reference. | PipelineReference (required) |
waitOnCompletion | Defines whether activity execution will wait for the dependent pipeline execution to finish. Default is false. | bool |
PipelineReference
Name | Description | Value |
---|---|---|
name | Reference name. | string |
referenceName | Reference pipeline name. | string (required) |
type | Pipeline reference type. | "PipelineReference" (required) |
ExecuteSsisPackageActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "ExecuteSSISPackage" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute SSIS package activity properties. | ExecuteSsisPackageActivityTypeProperties (required) |
ExecuteSsisPackageActivityTypeProperties
Name | Description | Value |
---|---|---|
connectVia | The integration runtime reference. | IntegrationRuntimeReference (required) |
environmentPath | The environment path to execute the SSIS package. Type: string (or Expression with resultType string). | |
executionCredential | The package execution credential. | SsisExecutionCredential |
loggingLevel | The logging level of SSIS package execution. Type: string (or Expression with resultType string). | |
logLocation | SSIS package execution log location. | SsisLogLocation |
packageConnectionManagers | The package level connection managers to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesPackageConne... |
packageLocation | SSIS package location. | SsisPackageLocation (required) |
packageParameters | The package level parameters to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesPackageParam... |
projectConnectionManagers | The project level connection managers to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesProjectConne... |
projectParameters | The project level parameters to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesProjectParam... |
propertyOverrides | The property overrides to execute the SSIS package. | ExecuteSsisPackageActivityTypePropertiesPropertyOver... |
runtime | Specifies the runtime to execute SSIS package. The value should be "x86" or "x64". Type: string (or Expression with resultType string). |
SsisExecutionCredential
Name | Description | Value |
---|---|---|
domain | Domain for windows authentication. Type: string (or Expression with resultType string). | |
password | Password for windows authentication. | SecureString (required) |
userName | UseName for windows authentication. Type: string (or Expression with resultType string). |
SecureString
Name | Description | Value |
---|---|---|
type | Type of the secret. | string (required) |
value | Value of secure string. | string (required) |
SsisLogLocation
Name | Description | Value |
---|---|---|
logPath | The SSIS package execution log path. Type: string (or Expression with resultType string). | |
type | The type of SSIS log location. | "File" (required) |
typeProperties | SSIS package execution log location properties. | SsisLogLocationTypeProperties (required) |
SsisLogLocationTypeProperties
Name | Description | Value |
---|---|---|
accessCredential | The package execution log access credential. | SsisAccessCredential |
logRefreshInterval | Specifies the interval to refresh log. The default interval is 5 minutes. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
SsisAccessCredential
Name | Description | Value |
---|---|---|
domain | Domain for windows authentication. Type: string (or Expression with resultType string). | |
password | Password for windows authentication. | SecretBase (required) |
userName | UseName for windows authentication. Type: string (or Expression with resultType string). |
SecretBase
Name | Description | Value |
---|---|---|
type | Set the object type | AzureKeyVaultSecret SecureString (required) |
AzureKeyVaultSecretReference
Name | Description | Value |
---|---|---|
secretName | The name of the secret in Azure Key Vault. Type: string (or Expression with resultType string). | |
secretVersion | The version of the secret in Azure Key Vault. The default value is the latest version of the secret. Type: string (or Expression with resultType string). | |
store | The Azure Key Vault linked service reference. | LinkedServiceReference (required) |
type | Type of the secret. | string (required) |
ExecuteSsisPackageActivityTypePropertiesPackageConne...
Name | Description | Value |
---|---|---|
{customized property} | SsisConnectionManager |
SsisConnectionManager
Name | Description | Value |
---|---|---|
{customized property} | SsisExecutionParameter |
SsisExecutionParameter
Name | Description | Value |
---|---|---|
value | SSIS package execution parameter value. Type: string (or Expression with resultType string). |
SsisPackageLocation
Name | Description | Value |
---|---|---|
packagePath | The SSIS package path. Type: string (or Expression with resultType string). | |
type | The type of SSIS package location. | "File" "InlinePackage" "PackageStore" "SSISDB" |
typeProperties | SSIS package location properties. | SsisPackageLocationTypeProperties |
SsisPackageLocationTypeProperties
Name | Description | Value |
---|---|---|
accessCredential | The package access credential. | SsisAccessCredential |
childPackages | The embedded child package list. | SsisChildPackage[] |
configurationAccessCredential | The configuration file access credential. | SsisAccessCredential |
configurationPath | The configuration file of the package execution. Type: string (or Expression with resultType string). | |
packageContent | The embedded package content. Type: string (or Expression with resultType string). | |
packageLastModifiedDate | The embedded package last modified date. | string |
packageName | The package name. | string |
packagePassword | Password of the package. | SecretBase |
SsisChildPackage
Name | Description | Value |
---|---|---|
packageContent | Content for embedded child package. Type: string (or Expression with resultType string). | |
packageLastModifiedDate | Last modified date for embedded child package. | string |
packageName | Name for embedded child package. | string |
packagePath | Path for embedded child package. Type: string (or Expression with resultType string). |
ExecuteSsisPackageActivityTypePropertiesPackageParam...
Name | Description | Value |
---|---|---|
{customized property} | SsisExecutionParameter |
ExecuteSsisPackageActivityTypePropertiesProjectConne...
Name | Description | Value |
---|---|---|
{customized property} | SsisConnectionManager |
ExecuteSsisPackageActivityTypePropertiesProjectParam...
Name | Description | Value |
---|---|---|
{customized property} | SsisExecutionParameter |
ExecuteSsisPackageActivityTypePropertiesPropertyOver...
Name | Description | Value |
---|---|---|
{customized property} | SsisPropertyOverride |
SsisPropertyOverride
Name | Description | Value |
---|---|---|
isSensitive | Whether SSIS package property override value is sensitive data. Value will be encrypted in SSISDB if it is true | bool |
value | SSIS package property override value. Type: string (or Expression with resultType string). |
ExecuteWranglingDataflowActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "ExecuteWranglingDataflow" (required) |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute power query activity properties. | ExecutePowerQueryActivityTypeProperties (required) |
ExecutePowerQueryActivityTypeProperties
Name | Description | Value |
---|---|---|
compute | Compute properties for data flow activity. | ExecuteDataFlowActivityTypePropertiesCompute |
continueOnError | Continue on error setting used for data flow execution. Enables processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean) | |
dataFlow | Data flow reference. | DataFlowReference (required) |
integrationRuntime | The integration runtime reference. | IntegrationRuntimeReference |
queries | List of mapping for Power Query mashup query to sink dataset(s). | PowerQuerySinkMapping[] |
runConcurrently | Concurrent run setting used for data flow execution. Allows sinks with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean) | |
sinks | (Deprecated. Please use Queries). List of Power Query activity sinks mapped to a queryName. | ExecutePowerQueryActivityTypePropertiesSinks |
sourceStagingConcurrency | Specify number of parallel staging for sources applicable to the sink. Type: integer (or Expression with resultType integer) | |
staging | Staging info for execute data flow activity. | DataFlowStagingInfo |
traceLevel | Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string) |
PowerQuerySinkMapping
Name | Description | Value |
---|---|---|
dataflowSinks | List of sinks mapped to Power Query mashup query. | PowerQuerySink[] |
queryName | Name of the query in Power Query mashup document. | string |
PowerQuerySink
Name | Description | Value |
---|---|---|
dataset | Dataset reference. | DatasetReference |
description | Transformation description. | string |
flowlet | Flowlet Reference | DataFlowReference |
linkedService | Linked service reference. | LinkedServiceReference |
name | Transformation name. | string (required) |
rejectedDataLinkedService | Rejected data linked service reference. | LinkedServiceReference |
schemaLinkedService | Schema linked service reference. | LinkedServiceReference |
script | sink script. | string |
ExecutePowerQueryActivityTypePropertiesSinks
Name | Description | Value |
---|---|---|
{customized property} | PowerQuerySink |
FailActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "Fail" (required) |
typeProperties | Fail activity properties. | FailActivityTypeProperties (required) |
FailActivityTypeProperties
Name | Description | Value |
---|---|---|
errorCode | The error code that categorizes the error type of the Fail activity. It can be dynamic content that's evaluated to a non empty/blank string at runtime. Type: string (or Expression with resultType string). | |
message | The error message that surfaced in the Fail activity. It can be dynamic content that's evaluated to a non empty/blank string at runtime. Type: string (or Expression with resultType string). |
FilterActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "Filter" (required) |
typeProperties | Filter activity properties. | FilterActivityTypeProperties (required) |
FilterActivityTypeProperties
Name | Description | Value |
---|---|---|
condition | Condition to be used for filtering the input. | Expression (required) |
items | Input array on which filter should be applied. | Expression (required) |
Expression
Name | Description | Value |
---|---|---|
type | Expression type. | "Expression" (required) |
value | Expression value. | string (required) |
ForEachActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "ForEach" (required) |
typeProperties | ForEach activity properties. | ForEachActivityTypeProperties (required) |
ForEachActivityTypeProperties
Name | Description | Value |
---|---|---|
activities | List of activities to execute . | Activity[] (required) |
batchCount | Batch count to be used for controlling the number of parallel execution (when isSequential is set to false). | int Constraints: Max value = 50 |
isSequential | Should the loop be executed in sequence or in parallel (max 50) | bool |
items | Collection to iterate. | Expression (required) |
GetMetadataActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "GetMetadata" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | GetMetadata activity properties. | GetMetadataActivityTypeProperties (required) |
GetMetadataActivityTypeProperties
Name | Description | Value |
---|---|---|
dataset | GetMetadata activity dataset reference. | DatasetReference (required) |
fieldList | Fields of metadata to get from dataset. | any[] |
formatSettings | GetMetadata activity format settings. | FormatReadSettings |
storeSettings | GetMetadata activity store settings. | StoreReadSettings |
FormatReadSettings
Name | Description | Value |
---|---|---|
type | Set the object type | BinaryReadSettings DelimitedTextReadSettings JsonReadSettings ParquetReadSettings XmlReadSettings (required) |
HDInsightHiveActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "HDInsightHive" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight Hive activity properties. | HDInsightHiveActivityTypeProperties (required) |
HDInsightHiveActivityTypeProperties
Name | Description | Value |
---|---|---|
arguments | User specified arguments to HDInsightActivity. | any[] |
defines | Allows user to specify defines for Hive job request. | HDInsightHiveActivityTypePropertiesDefines |
getDebugInfo | Debug info option. | "Always" "Failure" "None" |
queryTimeout | Query timeout value (in minutes). Effective when the HDInsight cluster is with ESP (Enterprise Security Package) | int |
scriptLinkedService | Script linked service reference. | LinkedServiceReference |
scriptPath | Script path. Type: string (or Expression with resultType string). | |
storageLinkedServices | Storage linked service references. | LinkedServiceReference[] |
variables | User specified arguments under hivevar namespace. | HDInsightHiveActivityTypePropertiesVariables |
HDInsightHiveActivityTypePropertiesDefines
Name | Description | Value |
---|---|---|
{customized property} |
HDInsightHiveActivityTypePropertiesVariables
Name | Description | Value |
---|---|---|
{customized property} |
HDInsightMapReduceActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "HDInsightMapReduce" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight MapReduce activity properties. | HDInsightMapReduceActivityTypeProperties (required) |
HDInsightMapReduceActivityTypeProperties
Name | Description | Value |
---|---|---|
arguments | User specified arguments to HDInsightActivity. | any[] |
className | Class name. Type: string (or Expression with resultType string). | |
defines | Allows user to specify defines for the MapReduce job request. | HDInsightMapReduceActivityTypePropertiesDefines |
getDebugInfo | Debug info option. | "Always" "Failure" "None" |
jarFilePath | Jar path. Type: string (or Expression with resultType string). | |
jarLibs | Jar libs. | any[] |
jarLinkedService | Jar linked service reference. | LinkedServiceReference |
storageLinkedServices | Storage linked service references. | LinkedServiceReference[] |
HDInsightMapReduceActivityTypePropertiesDefines
Name | Description | Value |
---|---|---|
{customized property} |
HDInsightPigActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "HDInsightPig" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight Pig activity properties. | HDInsightPigActivityTypeProperties (required) |
HDInsightPigActivityTypeProperties
Name | Description | Value |
---|---|---|
arguments | User specified arguments to HDInsightActivity. Type: array (or Expression with resultType array). | |
defines | Allows user to specify defines for Pig job request. | HDInsightPigActivityTypePropertiesDefines |
getDebugInfo | Debug info option. | "Always" "Failure" "None" |
scriptLinkedService | Script linked service reference. | LinkedServiceReference |
scriptPath | Script path. Type: string (or Expression with resultType string). | |
storageLinkedServices | Storage linked service references. | LinkedServiceReference[] |
HDInsightPigActivityTypePropertiesDefines
Name | Description | Value |
---|---|---|
{customized property} |
HDInsightSparkActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "HDInsightSpark" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight spark activity properties. | HDInsightSparkActivityTypeProperties (required) |
HDInsightSparkActivityTypeProperties
Name | Description | Value |
---|---|---|
arguments | The user-specified arguments to HDInsightSparkActivity. | any[] |
className | The application's Java/Spark main class. | string |
entryFilePath | The relative path to the root folder of the code/package to be executed. Type: string (or Expression with resultType string). | |
getDebugInfo | Debug info option. | "Always" "Failure" "None" |
proxyUser | The user to impersonate that will execute the job. Type: string (or Expression with resultType string). | |
rootPath | The root path in 'sparkJobLinkedService' for all the job’s files. Type: string (or Expression with resultType string). | |
sparkConfig | Spark configuration property. | HDInsightSparkActivityTypePropertiesSparkConfig |
sparkJobLinkedService | The storage linked service for uploading the entry file and dependencies, and for receiving logs. | LinkedServiceReference |
HDInsightSparkActivityTypePropertiesSparkConfig
Name | Description | Value |
---|---|---|
{customized property} |
HDInsightStreamingActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "HDInsightStreaming" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | HDInsight streaming activity properties. | HDInsightStreamingActivityTypeProperties (required) |
HDInsightStreamingActivityTypeProperties
Name | Description | Value |
---|---|---|
arguments | User specified arguments to HDInsightActivity. | any[] |
combiner | Combiner executable name. Type: string (or Expression with resultType string). | |
commandEnvironment | Command line environment values. | any[] |
defines | Allows user to specify defines for streaming job request. | HDInsightStreamingActivityTypePropertiesDefines |
fileLinkedService | Linked service reference where the files are located. | LinkedServiceReference |
filePaths | Paths to streaming job files. Can be directories. | any[] (required) |
getDebugInfo | Debug info option. | "Always" "Failure" "None" |
input | Input blob path. Type: string (or Expression with resultType string). | |
mapper | Mapper executable name. Type: string (or Expression with resultType string). | |
output | Output blob path. Type: string (or Expression with resultType string). | |
reducer | Reducer executable name. Type: string (or Expression with resultType string). | |
storageLinkedServices | Storage linked service references. | LinkedServiceReference[] |
HDInsightStreamingActivityTypePropertiesDefines
Name | Description | Value |
---|---|---|
{customized property} |
IfConditionActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "IfCondition" (required) |
typeProperties | IfCondition activity properties. | IfConditionActivityTypeProperties (required) |
IfConditionActivityTypeProperties
Name | Description | Value |
---|---|---|
expression | An expression that would evaluate to Boolean. This is used to determine the block of activities (ifTrueActivities or ifFalseActivities) that will be executed. | Expression (required) |
ifFalseActivities | List of activities to execute if expression is evaluated to false. This is an optional property and if not provided, the activity will exit without any action. | Activity[] |
ifTrueActivities | List of activities to execute if expression is evaluated to true. This is an optional property and if not provided, the activity will exit without any action. | Activity[] |
LookupActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "Lookup" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Lookup activity properties. | LookupActivityTypeProperties (required) |
LookupActivityTypeProperties
Name | Description | Value |
---|---|---|
dataset | Lookup activity dataset reference. | DatasetReference (required) |
firstRowOnly | Whether to return first row or all rows. Default value is true. Type: boolean (or Expression with resultType boolean). | |
source | Dataset-specific source properties, same as copy activity source. | CopySource (required) |
ScriptActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "Script" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Script activity properties. | ScriptActivityTypeProperties (required) |
ScriptActivityTypeProperties
Name | Description | Value |
---|---|---|
logSettings | Log settings of script activity. | ScriptActivityTypePropertiesLogSettings |
scriptBlockExecutionTimeout | ScriptBlock execution timeout. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | |
scripts | Array of script blocks. Type: array. | ScriptActivityScriptBlock[] |
ScriptActivityTypePropertiesLogSettings
Name | Description | Value |
---|---|---|
logDestination | The destination of logs. Type: string. | "ActivityOutput" "ExternalStore" (required) |
logLocationSettings | Log location settings customer needs to provide when enabling log. | LogLocationSettings |
ScriptActivityScriptBlock
Name | Description | Value |
---|---|---|
parameters | Array of script parameters. Type: array. | ScriptActivityParameter[] |
text | The query text. Type: string (or Expression with resultType string). | |
type | The type of the query. Type: string. | "NonQuery" "Query" (required) |
ScriptActivityParameter
Name | Description | Value |
---|---|---|
direction | The direction of the parameter. | "Input" "InputOutput" "Output" |
name | The name of the parameter. Type: string (or Expression with resultType string). | |
size | The size of the output direction parameter. | int |
type | The type of the parameter. | "Boolean" "DateTime" "DateTimeOffset" "Decimal" "Double" "Guid" "Int16" "Int32" "Int64" "Single" "String" "Timespan" |
value | The value of the parameter. Type: string (or Expression with resultType string). |
SetVariableActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "SetVariable" (required) |
policy | Activity policy. | SecureInputOutputPolicy |
typeProperties | Set Variable activity properties. | SetVariableActivityTypeProperties (required) |
SecureInputOutputPolicy
Name | Description | Value |
---|---|---|
secureInput | When set to true, Input from activity is considered as secure and will not be logged to monitoring. | bool |
secureOutput | When set to true, Output from activity is considered as secure and will not be logged to monitoring. | bool |
SetVariableActivityTypeProperties
Name | Description | Value |
---|---|---|
setSystemVariable | If set to true, it sets the pipeline run return value. | bool |
value | Value to be set. Could be a static value or Expression. | |
variableName | Name of the variable whose value needs to be set. | string |
SynapseSparkJobDefinitionActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "SparkJob" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute spark job activity properties. | SynapseSparkJobActivityTypeProperties (required) |
SynapseSparkJobActivityTypeProperties
Name | Description | Value |
---|---|---|
args | User specified arguments to SynapseSparkJobDefinitionActivity. | any[] |
className | The fully-qualified identifier or the main class that is in the main definition file, which will override the 'className' of the spark job definition you provide. Type: string (or Expression with resultType string). | |
conf | Spark configuration properties, which will override the 'conf' of the spark job definition you provide. | |
configurationType | The type of the spark config. | "Artifact" "Customized" "Default" |
driverSize | Number of core and memory to be used for driver allocated in the specified Spark pool for the job, which will be used for overriding 'driverCores' and 'driverMemory' of the spark job definition you provide. Type: string (or Expression with resultType string). | |
executorSize | Number of core and memory to be used for executors allocated in the specified Spark pool for the job, which will be used for overriding 'executorCores' and 'executorMemory' of the spark job definition you provide. Type: string (or Expression with resultType string). | |
file | The main file used for the job, which will override the 'file' of the spark job definition you provide. Type: string (or Expression with resultType string). | |
files | (Deprecated. Please use pythonCodeReference and filesV2) Additional files used for reference in the main definition file, which will override the 'files' of the spark job definition you provide. | any[] |
filesV2 | Additional files used for reference in the main definition file, which will override the 'jars' and 'files' of the spark job definition you provide. | any[] |
numExecutors | Number of executors to launch for this job, which will override the 'numExecutors' of the spark job definition you provide. Type: integer (or Expression with resultType integer). | |
pythonCodeReference | Additional python code files used for reference in the main definition file, which will override the 'pyFiles' of the spark job definition you provide. | any[] |
scanFolder | Scanning subfolders from the root folder of the main definition file, these files will be added as reference files. The folders named 'jars', 'pyFiles', 'files' or 'archives' will be scanned, and the folders name are case sensitive. Type: boolean (or Expression with resultType boolean). | |
sparkConfig | Spark configuration property. | SynapseSparkJobActivityTypePropertiesSparkConfig |
sparkJob | Synapse spark job reference. | SynapseSparkJobReference (required) |
targetBigDataPool | The name of the big data pool which will be used to execute the spark batch job, which will override the 'targetBigDataPool' of the spark job definition you provide. | BigDataPoolParametrizationReference |
targetSparkConfiguration | The spark configuration of the spark job. | SparkConfigurationParametrizationReference |
SynapseSparkJobActivityTypePropertiesSparkConfig
Name | Description | Value |
---|---|---|
{customized property} |
SynapseSparkJobReference
Name | Description | Value |
---|---|---|
referenceName | Reference spark job name. Expression with resultType string. | |
type | Synapse spark job reference type. | "SparkJobDefinitionReference" (required) |
BigDataPoolParametrizationReference
Name | Description | Value |
---|---|---|
referenceName | Reference big data pool name. Type: string (or Expression with resultType string). | |
type | Big data pool reference type. | "BigDataPoolReference" (required) |
SparkConfigurationParametrizationReference
Name | Description | Value |
---|---|---|
referenceName | Reference spark configuration name. Type: string (or Expression with resultType string). | |
type | Spark configuration reference type. | "SparkConfigurationReference" (required) |
SqlServerStoredProcedureActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "SqlServerStoredProcedure" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | SQL stored procedure activity properties. | SqlServerStoredProcedureActivityTypeProperties (required) |
SqlServerStoredProcedureActivityTypeProperties
Name | Description | Value |
---|---|---|
storedProcedureName | Stored procedure name. Type: string (or Expression with resultType string). | |
storedProcedureParameters | Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". |
SwitchActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "Switch" (required) |
typeProperties | Switch activity properties. | SwitchActivityTypeProperties (required) |
SwitchActivityTypeProperties
Name | Description | Value |
---|---|---|
cases | List of cases that correspond to expected values of the 'on' property. This is an optional property and if not provided, the activity will execute activities provided in defaultActivities. | SwitchCase[] |
defaultActivities | List of activities to execute if no case condition is satisfied. This is an optional property and if not provided, the activity will exit without any action. | Activity[] |
on | An expression that would evaluate to a string or integer. This is used to determine the block of activities in cases that will be executed. | Expression (required) |
SwitchCase
Name | Description | Value |
---|---|---|
activities | List of activities to execute for satisfied case condition. | Activity[] |
value | Expected value that satisfies the expression result of the 'on' property. | string |
SynapseNotebookActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "SynapseNotebook" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Execute Synapse notebook activity properties. | SynapseNotebookActivityTypeProperties (required) |
SynapseNotebookActivityTypeProperties
Name | Description | Value |
---|---|---|
conf | Spark configuration properties, which will override the 'conf' of the notebook you provide. | |
configurationType | The type of the spark config. | "Artifact" "Customized" "Default" |
driverSize | Number of core and memory to be used for driver allocated in the specified Spark pool for the session, which will be used for overriding 'driverCores' and 'driverMemory' of the notebook you provide. Type: string (or Expression with resultType string). | |
executorSize | Number of core and memory to be used for executors allocated in the specified Spark pool for the session, which will be used for overriding 'executorCores' and 'executorMemory' of the notebook you provide. Type: string (or Expression with resultType string). | |
notebook | Synapse notebook reference. | SynapseNotebookReference (required) |
numExecutors | Number of executors to launch for this session, which will override the 'numExecutors' of the notebook you provide. Type: integer (or Expression with resultType integer). | |
parameters | Notebook parameters. | SynapseNotebookActivityTypePropertiesParameters |
sparkConfig | Spark configuration property. | SynapseNotebookActivityTypePropertiesSparkConfig |
sparkPool | The name of the big data pool which will be used to execute the notebook. | BigDataPoolParametrizationReference |
targetSparkConfiguration | The spark configuration of the spark job. | SparkConfigurationParametrizationReference |
SynapseNotebookReference
Name | Description | Value |
---|---|---|
referenceName | Reference notebook name. Type: string (or Expression with resultType string). | |
type | Synapse notebook reference type. | "NotebookReference" (required) |
SynapseNotebookActivityTypePropertiesParameters
Name | Description | Value |
---|---|---|
{customized property} | NotebookParameter |
NotebookParameter
Name | Description | Value |
---|---|---|
type | Notebook parameter type. | "bool" "float" "int" "string" |
value | Notebook parameter value. Type: string (or Expression with resultType string). |
SynapseNotebookActivityTypePropertiesSparkConfig
Name | Description | Value |
---|---|---|
{customized property} |
UntilActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "Until" (required) |
typeProperties | Until activity properties. | UntilActivityTypeProperties (required) |
UntilActivityTypeProperties
Name | Description | Value |
---|---|---|
activities | List of activities to execute. | Activity[] (required) |
expression | An expression that would evaluate to Boolean. The loop will continue until this expression evaluates to true | Expression (required) |
timeout | Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
ValidationActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "Validation" (required) |
typeProperties | Validation activity properties. | ValidationActivityTypeProperties (required) |
ValidationActivityTypeProperties
Name | Description | Value |
---|---|---|
childItems | Can be used if dataset points to a folder. If set to true, the folder must have at least one file. If set to false, the folder must be empty. Type: boolean (or Expression with resultType boolean). | |
dataset | Validation activity dataset reference. | DatasetReference (required) |
minimumSize | Can be used if dataset points to a file. The file must be greater than or equal in size to the value specified. Type: integer (or Expression with resultType integer). | |
sleep | A delay in seconds between validation attempts. If no value is specified, 10 seconds will be used as the default. Type: integer (or Expression with resultType integer). | |
timeout | Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). |
WaitActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "Wait" (required) |
typeProperties | Wait activity properties. | WaitActivityTypeProperties (required) |
WaitActivityTypeProperties
Name | Description | Value |
---|---|---|
waitTimeInSeconds | Duration in seconds. Type: integer (or Expression with resultType integer). |
WebActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "WebActivity" (required) |
linkedServiceName | Linked service reference. | LinkedServiceReference |
policy | Activity policy. | ActivityPolicy |
typeProperties | Web activity properties. | WebActivityTypeProperties (required) |
WebActivityTypeProperties
Name | Description | Value |
---|---|---|
authentication | Authentication method used for calling the endpoint. | WebActivityAuthentication |
body | Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). | |
connectVia | The integration runtime reference. | IntegrationRuntimeReference |
datasets | List of datasets passed to web endpoint. | DatasetReference[] |
disableCertValidation | When set to true, Certificate validation will be disabled. | bool |
headers | Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). | WebActivityTypePropertiesHeaders |
httpRequestTimeout | Timeout for the HTTP request to get a response. Format is in TimeSpan (hh:mm:ss). This value is the timeout to get a response, not the activity timeout. The default value is 00:01:00 (1 minute). The range is from 1 to 10 minutes | |
linkedServices | List of linked services passed to web endpoint. | LinkedServiceReference[] |
method | Rest API method for target endpoint. | "DELETE" "GET" "POST" "PUT" (required) |
turnOffAsync | Option to disable invoking HTTP GET on location given in response header of a HTTP 202 Response. If set true, it stops invoking HTTP GET on http location given in response header. If set false then continues to invoke HTTP GET call on location given in http response headers. | bool |
url | Web activity target endpoint and path. Type: string (or Expression with resultType string). |
WebActivityAuthentication
Name | Description | Value |
---|---|---|
credential | The credential reference containing authentication information. | CredentialReference |
password | Password for the PFX file or basic authentication / Secret when used for ServicePrincipal | SecretBase |
pfx | Base64-encoded contents of a PFX file or Certificate when used for ServicePrincipal | SecretBase |
resource | Resource for which Azure Auth token will be requested when using MSI Authentication. Type: string (or Expression with resultType string). | |
type | Web activity authentication (Basic/ClientCertificate/MSI/ServicePrincipal) | string |
username | Web activity authentication user name for basic authentication or ClientID when used for ServicePrincipal. Type: string (or Expression with resultType string). | |
userTenant | TenantId for which Azure Auth token will be requested when using ServicePrincipal Authentication. Type: string (or Expression with resultType string). |
CredentialReference
Name | Description | Value |
---|---|---|
referenceName | Reference credential name. | string (required) |
type | Credential reference type. | "CredentialReference" (required) |
{customized property} |
WebActivityTypePropertiesHeaders
Name | Description | Value |
---|---|---|
{customized property} | string |
WebHookActivity
Name | Description | Value |
---|---|---|
type | Type of activity. | "WebHook" (required) |
policy | Activity policy. | SecureInputOutputPolicy |
typeProperties | WebHook activity properties. | WebHookActivityTypeProperties (required) |
WebHookActivityTypeProperties
Name | Description | Value |
---|---|---|
authentication | Authentication method used for calling the endpoint. | WebActivityAuthentication |
body | Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). | |
headers | Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": "application/json" }. Type: string (or Expression with resultType string). | WebHookActivityTypePropertiesHeaders |
method | Rest API method for target endpoint. | "POST" (required) |
reportStatusOnCallBack | When set to true, statusCode, output and error in callback request body will be consumed by activity. The activity can be marked as failed by setting statusCode >= 400 in callback request. Default is false. Type: boolean (or Expression with resultType boolean). | |
timeout | The timeout within which the webhook should be called back. If there is no value specified, it defaults to 10 minutes. Type: string. Pattern: ((\d+).)?(\d\d):(60,([0-5][0-9])):(60,([0-5][0-9])). | string |
url | WebHook activity target endpoint and path. Type: string (or Expression with resultType string). |
WebHookActivityTypePropertiesHeaders
Name | Description | Value |
---|---|---|
{customized property} | string |
PipelineFolder
Name | Description | Value |
---|---|---|
name | The name of the folder that this Pipeline is in. | string |
ParameterDefinitionSpecification
Name | Description | Value |
---|---|---|
{customized property} | ParameterSpecification |
ParameterSpecification
Name | Description | Value |
---|---|---|
defaultValue | Default value of parameter. | |
type | Parameter type. | "Array" "Bool" "Float" "Int" "Object" "SecureString" "String" (required) |
PipelinePolicy
Name | Description | Value |
---|---|---|
elapsedTimeMetric | Pipeline ElapsedTime Metric Policy. | PipelineElapsedTimeMetricPolicy |
PipelineElapsedTimeMetricPolicy
Name | Description | Value |
---|---|---|
duration | TimeSpan value, after which an Azure Monitoring Metric is fired. |
PipelineRunDimensions
Name | Description | Value |
---|---|---|
{customized property} |
VariableDefinitionSpecification
Name | Description | Value |
---|---|---|
{customized property} | VariableSpecification |
VariableSpecification
Name | Description | Value |
---|---|---|
defaultValue | Default value of variable. | |
type | Variable type. | "Array" "Bool" "String" (required) |