diff --git a/sdk/synapse/azure-synapse-artifacts/CHANGELOG.md b/sdk/synapse/azure-synapse-artifacts/CHANGELOG.md index 253ee6252d7b..644dc02814a0 100644 --- a/sdk/synapse/azure-synapse-artifacts/CHANGELOG.md +++ b/sdk/synapse/azure-synapse-artifacts/CHANGELOG.md @@ -1,14 +1,10 @@ # Release History -## 0.9.0 (Unreleased) +## 0.9.0 (2021-10-05) ### Features Added -### Breaking Changes - -### Bugs Fixed - -### Other Changes +- re-generated based on tag package-artifacts-composite-v1 ## 0.8.0 (2021-08-10) diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/__init__.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/__init__.py index 0a8d3bc734b7..03703237ffc5 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/__init__.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/__init__.py @@ -7,6 +7,9 @@ # -------------------------------------------------------------------------- from ._artifacts_client import ArtifactsClient +from ._version import VERSION + +__version__ = VERSION __all__ = ['ArtifactsClient'] try: diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/_artifacts_client.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/_artifacts_client.py index 4f589ed2ef42..9eedec5137df 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/_artifacts_client.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/_artifacts_client.py @@ -1,443 +1,186 @@ # coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# +# Licensed under the MIT License. See License.txt in the project root for license information. # Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- from typing import TYPE_CHECKING from azure.core import PipelineClient -from azure.profiles import KnownProfiles, ProfileDefinition -from azure.profiles.multiapiclient import MultiApiClientMixin from msrest import Deserializer, Serializer -from ._configuration import ArtifactsClientConfiguration - if TYPE_CHECKING: # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Optional + from typing import Any from azure.core.credentials import TokenCredential from azure.core.pipeline.transport import HttpRequest, HttpResponse -class _SDKClient(object): - def __init__(self, *args, **kwargs): - """This is a fake class to support current implemetation of MultiApiClientMixin." - Will be removed in final version of multiapi azure-core based client - """ - pass - -class ArtifactsClient(MultiApiClientMixin, _SDKClient): +from ._configuration import ArtifactsClientConfiguration +from .operations import KqlScriptsOperations +from .operations import KqlScriptOperations +from .operations import SparkConfigurationOperations +from .operations import BigDataPoolsOperations +from .operations import DataFlowOperations +from .operations import DataFlowDebugSessionOperations +from .operations import DatasetOperations +from .operations import WorkspaceGitRepoManagementOperations +from .operations import IntegrationRuntimesOperations +from .operations import LibraryOperations +from .operations import LinkedServiceOperations +from .operations import NotebookOperations +from .operations import NotebookOperationResultOperations +from .operations import PipelineOperations +from .operations import PipelineRunOperations +from .operations import SparkJobDefinitionOperations +from .operations import SqlPoolsOperations +from .operations import SqlScriptOperations +from .operations import TriggerOperations +from .operations import TriggerRunOperations +from .operations import WorkspaceOperations +from . import models + + +class ArtifactsClient(object): """ArtifactsClient. - This ready contains multiple API versions, to help you deal with all of the Azure clouds - (Azure Stack, Azure Government, Azure China, etc.). - By default, it uses the latest API version available on public Azure. - For production, you should stick to a particular api-version and/or profile. - The profile sets a mapping between an operation group and its API version. - The api-version parameter sets the default API version if the operation - group is not described in the profile. - + :ivar kql_scripts: KqlScriptsOperations operations + :vartype kql_scripts: azure.synapse.artifacts.operations.KqlScriptsOperations + :ivar kql_script: KqlScriptOperations operations + :vartype kql_script: azure.synapse.artifacts.operations.KqlScriptOperations + :ivar spark_configuration: SparkConfigurationOperations operations + :vartype spark_configuration: azure.synapse.artifacts.operations.SparkConfigurationOperations + :ivar big_data_pools: BigDataPoolsOperations operations + :vartype big_data_pools: azure.synapse.artifacts.operations.BigDataPoolsOperations + :ivar data_flow: DataFlowOperations operations + :vartype data_flow: azure.synapse.artifacts.operations.DataFlowOperations + :ivar data_flow_debug_session: DataFlowDebugSessionOperations operations + :vartype data_flow_debug_session: azure.synapse.artifacts.operations.DataFlowDebugSessionOperations + :ivar dataset: DatasetOperations operations + :vartype dataset: azure.synapse.artifacts.operations.DatasetOperations + :ivar workspace_git_repo_management: WorkspaceGitRepoManagementOperations operations + :vartype workspace_git_repo_management: azure.synapse.artifacts.operations.WorkspaceGitRepoManagementOperations + :ivar integration_runtimes: IntegrationRuntimesOperations operations + :vartype integration_runtimes: azure.synapse.artifacts.operations.IntegrationRuntimesOperations + :ivar library: LibraryOperations operations + :vartype library: azure.synapse.artifacts.operations.LibraryOperations + :ivar linked_service: LinkedServiceOperations operations + :vartype linked_service: azure.synapse.artifacts.operations.LinkedServiceOperations + :ivar notebook: NotebookOperations operations + :vartype notebook: azure.synapse.artifacts.operations.NotebookOperations + :ivar notebook_operation_result: NotebookOperationResultOperations operations + :vartype notebook_operation_result: azure.synapse.artifacts.operations.NotebookOperationResultOperations + :ivar pipeline: PipelineOperations operations + :vartype pipeline: azure.synapse.artifacts.operations.PipelineOperations + :ivar pipeline_run: PipelineRunOperations operations + :vartype pipeline_run: azure.synapse.artifacts.operations.PipelineRunOperations + :ivar spark_job_definition: SparkJobDefinitionOperations operations + :vartype spark_job_definition: azure.synapse.artifacts.operations.SparkJobDefinitionOperations + :ivar sql_pools: SqlPoolsOperations operations + :vartype sql_pools: azure.synapse.artifacts.operations.SqlPoolsOperations + :ivar sql_script: SqlScriptOperations operations + :vartype sql_script: azure.synapse.artifacts.operations.SqlScriptOperations + :ivar trigger: TriggerOperations operations + :vartype trigger: azure.synapse.artifacts.operations.TriggerOperations + :ivar trigger_run: TriggerRunOperations operations + :vartype trigger_run: azure.synapse.artifacts.operations.TriggerRunOperations + :ivar workspace: WorkspaceOperations operations + :vartype workspace: azure.synapse.artifacts.operations.WorkspaceOperations :param credential: Credential needed for the client to connect to Azure. :type credential: ~azure.core.credentials.TokenCredential :param endpoint: The workspace development endpoint, for example https://myworkspace.dev.azuresynapse.net. :type endpoint: str - :param api_version: API version to use if no profile is provided, or if missing in profile. - :type api_version: str - :param profile: A profile definition, from KnownProfiles to dict. - :type profile: azure.profiles.KnownProfiles :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. """ - DEFAULT_API_VERSION = '2020-12-01' - _PROFILE_TAG = "azure.synapse.artifacts.ArtifactsClient" - LATEST_PROFILE = ProfileDefinition({ - _PROFILE_TAG: { - None: DEFAULT_API_VERSION, - }}, - _PROFILE_TAG + " latest" - ) - def __init__( self, credential, # type: "TokenCredential" endpoint, # type: str - api_version=DEFAULT_API_VERSION, # type: Optional[str] - profile=KnownProfiles.default, # type: KnownProfiles **kwargs # type: Any ): - if api_version == '2020-12-01' or api_version == '2021-06-01-preview': - base_url = '{endpoint}' - else: - raise ValueError("API version {} is not available".format(api_version)) + # type: (...) -> None + base_url = '{endpoint}' self._config = ArtifactsClientConfiguration(credential, endpoint, **kwargs) self._client = PipelineClient(base_url=base_url, config=self._config, **kwargs) - super(ArtifactsClient, self).__init__( - api_version=api_version, - profile=profile - ) - - @classmethod - def _models_dict(cls, api_version): - return {k: v for k, v in cls.models(api_version).__dict__.items() if isinstance(v, type)} - - @classmethod - def models(cls, api_version=DEFAULT_API_VERSION): - """Module depends on the API version: - - * 2020-12-01: :mod:`v2020_12_01.models` - * 2021-06-01-preview: :mod:`v2021_06_01_preview.models` - """ - if api_version == '2020-12-01': - from .v2020_12_01 import models - return models - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview import models - return models - raise ValueError("API version {} is not available".format(api_version)) - - @property - def big_data_pools(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`BigDataPoolsOperations` - * 2021-06-01-preview: :class:`BigDataPoolsOperations` - """ - api_version = self._get_api_version('big_data_pools') - if api_version == '2020-12-01': - from .v2020_12_01.operations import BigDataPoolsOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import BigDataPoolsOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'big_data_pools'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def data_flow(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`DataFlowOperations` - * 2021-06-01-preview: :class:`DataFlowOperations` - """ - api_version = self._get_api_version('data_flow') - if api_version == '2020-12-01': - from .v2020_12_01.operations import DataFlowOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import DataFlowOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'data_flow'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def data_flow_debug_session(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`DataFlowDebugSessionOperations` - * 2021-06-01-preview: :class:`DataFlowDebugSessionOperations` - """ - api_version = self._get_api_version('data_flow_debug_session') - if api_version == '2020-12-01': - from .v2020_12_01.operations import DataFlowDebugSessionOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import DataFlowDebugSessionOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'data_flow_debug_session'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def dataset(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`DatasetOperations` - * 2021-06-01-preview: :class:`DatasetOperations` - """ - api_version = self._get_api_version('dataset') - if api_version == '2020-12-01': - from .v2020_12_01.operations import DatasetOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import DatasetOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'dataset'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def integration_runtimes(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`IntegrationRuntimesOperations` - * 2021-06-01-preview: :class:`IntegrationRuntimesOperations` - """ - api_version = self._get_api_version('integration_runtimes') - if api_version == '2020-12-01': - from .v2020_12_01.operations import IntegrationRuntimesOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import IntegrationRuntimesOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'integration_runtimes'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def library(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`LibraryOperations` - * 2021-06-01-preview: :class:`LibraryOperations` - """ - api_version = self._get_api_version('library') - if api_version == '2020-12-01': - from .v2020_12_01.operations import LibraryOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import LibraryOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'library'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def linked_service(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`LinkedServiceOperations` - * 2021-06-01-preview: :class:`LinkedServiceOperations` - """ - api_version = self._get_api_version('linked_service') - if api_version == '2020-12-01': - from .v2020_12_01.operations import LinkedServiceOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import LinkedServiceOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'linked_service'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def notebook(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`NotebookOperations` - * 2021-06-01-preview: :class:`NotebookOperations` - """ - api_version = self._get_api_version('notebook') - if api_version == '2020-12-01': - from .v2020_12_01.operations import NotebookOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import NotebookOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'notebook'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def notebook_operation_result(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`NotebookOperationResultOperations` - * 2021-06-01-preview: :class:`NotebookOperationResultOperations` - """ - api_version = self._get_api_version('notebook_operation_result') - if api_version == '2020-12-01': - from .v2020_12_01.operations import NotebookOperationResultOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import NotebookOperationResultOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'notebook_operation_result'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - @property - def operation_result(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`OperationResultOperations` - * 2021-06-01-preview: :class:`OperationResultOperations` - """ - api_version = self._get_api_version('operation_result') - if api_version == '2020-12-01': - from .v2020_12_01.operations import OperationResultOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import OperationResultOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'operation_result'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def operation_status(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`OperationStatusOperations` - * 2021-06-01-preview: :class:`OperationStatusOperations` - """ - api_version = self._get_api_version('operation_status') - if api_version == '2020-12-01': - from .v2020_12_01.operations import OperationStatusOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import OperationStatusOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'operation_status'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def pipeline(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`PipelineOperations` - * 2021-06-01-preview: :class:`PipelineOperations` - """ - api_version = self._get_api_version('pipeline') - if api_version == '2020-12-01': - from .v2020_12_01.operations import PipelineOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import PipelineOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'pipeline'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def pipeline_run(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`PipelineRunOperations` - * 2021-06-01-preview: :class:`PipelineRunOperations` - """ - api_version = self._get_api_version('pipeline_run') - if api_version == '2020-12-01': - from .v2020_12_01.operations import PipelineRunOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import PipelineRunOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'pipeline_run'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def spark_configuration(self): - """Instance depends on the API version: - - * 2021-06-01-preview: :class:`SparkConfigurationOperations` - """ - api_version = self._get_api_version('spark_configuration') - if api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import SparkConfigurationOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'spark_configuration'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def spark_job_definition(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`SparkJobDefinitionOperations` - * 2021-06-01-preview: :class:`SparkJobDefinitionOperations` - """ - api_version = self._get_api_version('spark_job_definition') - if api_version == '2020-12-01': - from .v2020_12_01.operations import SparkJobDefinitionOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import SparkJobDefinitionOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'spark_job_definition'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def sql_pools(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`SqlPoolsOperations` - * 2021-06-01-preview: :class:`SqlPoolsOperations` - """ - api_version = self._get_api_version('sql_pools') - if api_version == '2020-12-01': - from .v2020_12_01.operations import SqlPoolsOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import SqlPoolsOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'sql_pools'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def sql_script(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`SqlScriptOperations` - * 2021-06-01-preview: :class:`SqlScriptOperations` - """ - api_version = self._get_api_version('sql_script') - if api_version == '2020-12-01': - from .v2020_12_01.operations import SqlScriptOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import SqlScriptOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'sql_script'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def trigger(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`TriggerOperations` - * 2021-06-01-preview: :class:`TriggerOperations` - """ - api_version = self._get_api_version('trigger') - if api_version == '2020-12-01': - from .v2020_12_01.operations import TriggerOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import TriggerOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'trigger'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def trigger_run(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`TriggerRunOperations` - * 2021-06-01-preview: :class:`TriggerRunOperations` - """ - api_version = self._get_api_version('trigger_run') - if api_version == '2020-12-01': - from .v2020_12_01.operations import TriggerRunOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import TriggerRunOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'trigger_run'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def workspace(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`WorkspaceOperations` - * 2021-06-01-preview: :class:`WorkspaceOperations` - """ - api_version = self._get_api_version('workspace') - if api_version == '2020-12-01': - from .v2020_12_01.operations import WorkspaceOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import WorkspaceOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'workspace'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def workspace_git_repo_management(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`WorkspaceGitRepoManagementOperations` - * 2021-06-01-preview: :class:`WorkspaceGitRepoManagementOperations` - """ - api_version = self._get_api_version('workspace_git_repo_management') - if api_version == '2020-12-01': - from .v2020_12_01.operations import WorkspaceGitRepoManagementOperations as OperationClass - elif api_version == '2021-06-01-preview': - from .v2021_06_01_preview.operations import WorkspaceGitRepoManagementOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'workspace_git_repo_management'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) + client_models = {k: v for k, v in models.__dict__.items() if isinstance(v, type)} + self._serialize = Serializer(client_models) + self._serialize.client_side_validation = False + self._deserialize = Deserializer(client_models) + + self.kql_scripts = KqlScriptsOperations( + self._client, self._config, self._serialize, self._deserialize) + self.kql_script = KqlScriptOperations( + self._client, self._config, self._serialize, self._deserialize) + self.spark_configuration = SparkConfigurationOperations( + self._client, self._config, self._serialize, self._deserialize) + self.big_data_pools = BigDataPoolsOperations( + self._client, self._config, self._serialize, self._deserialize) + self.data_flow = DataFlowOperations( + self._client, self._config, self._serialize, self._deserialize) + self.data_flow_debug_session = DataFlowDebugSessionOperations( + self._client, self._config, self._serialize, self._deserialize) + self.dataset = DatasetOperations( + self._client, self._config, self._serialize, self._deserialize) + self.workspace_git_repo_management = WorkspaceGitRepoManagementOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtimes = IntegrationRuntimesOperations( + self._client, self._config, self._serialize, self._deserialize) + self.library = LibraryOperations( + self._client, self._config, self._serialize, self._deserialize) + self.linked_service = LinkedServiceOperations( + self._client, self._config, self._serialize, self._deserialize) + self.notebook = NotebookOperations( + self._client, self._config, self._serialize, self._deserialize) + self.notebook_operation_result = NotebookOperationResultOperations( + self._client, self._config, self._serialize, self._deserialize) + self.pipeline = PipelineOperations( + self._client, self._config, self._serialize, self._deserialize) + self.pipeline_run = PipelineRunOperations( + self._client, self._config, self._serialize, self._deserialize) + self.spark_job_definition = SparkJobDefinitionOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pools = SqlPoolsOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_script = SqlScriptOperations( + self._client, self._config, self._serialize, self._deserialize) + self.trigger = TriggerOperations( + self._client, self._config, self._serialize, self._deserialize) + self.trigger_run = TriggerRunOperations( + self._client, self._config, self._serialize, self._deserialize) + self.workspace = WorkspaceOperations( + self._client, self._config, self._serialize, self._deserialize) + + def _send_request(self, http_request, **kwargs): + # type: (HttpRequest, Any) -> HttpResponse + """Runs the network request through the client's chained policies. + + :param http_request: The network request you want to make. Required. + :type http_request: ~azure.core.pipeline.transport.HttpRequest + :keyword bool stream: Whether the response payload will be streamed. Defaults to True. + :return: The response of your network call. Does not do error handling on your response. + :rtype: ~azure.core.pipeline.transport.HttpResponse + """ + path_format_arguments = { + 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), + } + http_request.url = self._client.format_url(http_request.url, **path_format_arguments) + stream = kwargs.pop("stream", True) + pipeline_response = self._client._pipeline.run(http_request, stream=stream, **kwargs) + return pipeline_response.http_response def close(self): + # type: () -> None self._client.close() + def __enter__(self): + # type: () -> ArtifactsClient self._client.__enter__() return self + def __exit__(self, *exc_details): + # type: (Any) -> None self._client.__exit__(*exc_details) diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/_configuration.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/_configuration.py index 879c776a1e19..eac5385e6a35 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/_configuration.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/_configuration.py @@ -1,13 +1,11 @@ # coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# +# Licensed under the MIT License. See License.txt in the project root for license information. # Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- + from typing import TYPE_CHECKING from azure.core.configuration import Configuration @@ -21,6 +19,7 @@ from azure.core.credentials import TokenCredential + class ArtifactsClientConfiguration(Configuration): """Configuration for ArtifactsClient. @@ -49,7 +48,7 @@ def __init__( self.credential = credential self.endpoint = endpoint self.credential_scopes = kwargs.pop('credential_scopes', ['https://dev.azuresynapse.net/.default']) - kwargs.setdefault('sdk_moniker', 'azure-synapse-artifacts/{}'.format(VERSION)) + kwargs.setdefault('sdk_moniker', 'synapse-artifacts/{}'.format(VERSION)) self._configure(**kwargs) def _configure( diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/_artifacts_client.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/_artifacts_client.py index c9a6a716a343..8e674570b855 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/_artifacts_client.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/_artifacts_client.py @@ -1,441 +1,179 @@ # coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# +# Licensed under the MIT License. See License.txt in the project root for license information. # Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from typing import Any, Optional, TYPE_CHECKING +from typing import Any, TYPE_CHECKING from azure.core import AsyncPipelineClient from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest -from azure.profiles import KnownProfiles, ProfileDefinition -from azure.profiles.multiapiclient import MultiApiClientMixin from msrest import Deserializer, Serializer -from ._configuration import ArtifactsClientConfiguration - if TYPE_CHECKING: # pylint: disable=unused-import,ungrouped-imports from azure.core.credentials_async import AsyncTokenCredential -class _SDKClient(object): - def __init__(self, *args, **kwargs): - """This is a fake class to support current implemetation of MultiApiClientMixin." - Will be removed in final version of multiapi azure-core based client - """ - pass - -class ArtifactsClient(MultiApiClientMixin, _SDKClient): +from ._configuration import ArtifactsClientConfiguration +from .operations import KqlScriptsOperations +from .operations import KqlScriptOperations +from .operations import SparkConfigurationOperations +from .operations import BigDataPoolsOperations +from .operations import DataFlowOperations +from .operations import DataFlowDebugSessionOperations +from .operations import DatasetOperations +from .operations import WorkspaceGitRepoManagementOperations +from .operations import IntegrationRuntimesOperations +from .operations import LibraryOperations +from .operations import LinkedServiceOperations +from .operations import NotebookOperations +from .operations import NotebookOperationResultOperations +from .operations import PipelineOperations +from .operations import PipelineRunOperations +from .operations import SparkJobDefinitionOperations +from .operations import SqlPoolsOperations +from .operations import SqlScriptOperations +from .operations import TriggerOperations +from .operations import TriggerRunOperations +from .operations import WorkspaceOperations +from .. import models + + +class ArtifactsClient(object): """ArtifactsClient. - This ready contains multiple API versions, to help you deal with all of the Azure clouds - (Azure Stack, Azure Government, Azure China, etc.). - By default, it uses the latest API version available on public Azure. - For production, you should stick to a particular api-version and/or profile. - The profile sets a mapping between an operation group and its API version. - The api-version parameter sets the default API version if the operation - group is not described in the profile. - + :ivar kql_scripts: KqlScriptsOperations operations + :vartype kql_scripts: azure.synapse.artifacts.aio.operations.KqlScriptsOperations + :ivar kql_script: KqlScriptOperations operations + :vartype kql_script: azure.synapse.artifacts.aio.operations.KqlScriptOperations + :ivar spark_configuration: SparkConfigurationOperations operations + :vartype spark_configuration: azure.synapse.artifacts.aio.operations.SparkConfigurationOperations + :ivar big_data_pools: BigDataPoolsOperations operations + :vartype big_data_pools: azure.synapse.artifacts.aio.operations.BigDataPoolsOperations + :ivar data_flow: DataFlowOperations operations + :vartype data_flow: azure.synapse.artifacts.aio.operations.DataFlowOperations + :ivar data_flow_debug_session: DataFlowDebugSessionOperations operations + :vartype data_flow_debug_session: azure.synapse.artifacts.aio.operations.DataFlowDebugSessionOperations + :ivar dataset: DatasetOperations operations + :vartype dataset: azure.synapse.artifacts.aio.operations.DatasetOperations + :ivar workspace_git_repo_management: WorkspaceGitRepoManagementOperations operations + :vartype workspace_git_repo_management: azure.synapse.artifacts.aio.operations.WorkspaceGitRepoManagementOperations + :ivar integration_runtimes: IntegrationRuntimesOperations operations + :vartype integration_runtimes: azure.synapse.artifacts.aio.operations.IntegrationRuntimesOperations + :ivar library: LibraryOperations operations + :vartype library: azure.synapse.artifacts.aio.operations.LibraryOperations + :ivar linked_service: LinkedServiceOperations operations + :vartype linked_service: azure.synapse.artifacts.aio.operations.LinkedServiceOperations + :ivar notebook: NotebookOperations operations + :vartype notebook: azure.synapse.artifacts.aio.operations.NotebookOperations + :ivar notebook_operation_result: NotebookOperationResultOperations operations + :vartype notebook_operation_result: azure.synapse.artifacts.aio.operations.NotebookOperationResultOperations + :ivar pipeline: PipelineOperations operations + :vartype pipeline: azure.synapse.artifacts.aio.operations.PipelineOperations + :ivar pipeline_run: PipelineRunOperations operations + :vartype pipeline_run: azure.synapse.artifacts.aio.operations.PipelineRunOperations + :ivar spark_job_definition: SparkJobDefinitionOperations operations + :vartype spark_job_definition: azure.synapse.artifacts.aio.operations.SparkJobDefinitionOperations + :ivar sql_pools: SqlPoolsOperations operations + :vartype sql_pools: azure.synapse.artifacts.aio.operations.SqlPoolsOperations + :ivar sql_script: SqlScriptOperations operations + :vartype sql_script: azure.synapse.artifacts.aio.operations.SqlScriptOperations + :ivar trigger: TriggerOperations operations + :vartype trigger: azure.synapse.artifacts.aio.operations.TriggerOperations + :ivar trigger_run: TriggerRunOperations operations + :vartype trigger_run: azure.synapse.artifacts.aio.operations.TriggerRunOperations + :ivar workspace: WorkspaceOperations operations + :vartype workspace: azure.synapse.artifacts.aio.operations.WorkspaceOperations :param credential: Credential needed for the client to connect to Azure. :type credential: ~azure.core.credentials_async.AsyncTokenCredential :param endpoint: The workspace development endpoint, for example https://myworkspace.dev.azuresynapse.net. :type endpoint: str - :param api_version: API version to use if no profile is provided, or if missing in profile. - :type api_version: str - :param profile: A profile definition, from KnownProfiles to dict. - :type profile: azure.profiles.KnownProfiles :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. """ - DEFAULT_API_VERSION = '2020-12-01' - _PROFILE_TAG = "azure.synapse.artifacts.ArtifactsClient" - LATEST_PROFILE = ProfileDefinition({ - _PROFILE_TAG: { - None: DEFAULT_API_VERSION, - }}, - _PROFILE_TAG + " latest" - ) - def __init__( self, credential: "AsyncTokenCredential", endpoint: str, - api_version: Optional[str] = DEFAULT_API_VERSION, - profile: KnownProfiles = KnownProfiles.default, - **kwargs # type: Any + **kwargs: Any ) -> None: - if api_version == '2020-12-01' or api_version == '2021-06-01-preview': - base_url = '{endpoint}' - else: - raise ValueError("API version {} is not available".format(api_version)) + base_url = '{endpoint}' self._config = ArtifactsClientConfiguration(credential, endpoint, **kwargs) self._client = AsyncPipelineClient(base_url=base_url, config=self._config, **kwargs) - super(ArtifactsClient, self).__init__( - api_version=api_version, - profile=profile - ) - - @classmethod - def _models_dict(cls, api_version): - return {k: v for k, v in cls.models(api_version).__dict__.items() if isinstance(v, type)} - - @classmethod - def models(cls, api_version=DEFAULT_API_VERSION): - """Module depends on the API version: - - * 2020-12-01: :mod:`v2020_12_01.models` - * 2021-06-01-preview: :mod:`v2021_06_01_preview.models` - """ - if api_version == '2020-12-01': - from ..v2020_12_01 import models - return models - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview import models - return models - raise ValueError("API version {} is not available".format(api_version)) - - @property - def big_data_pools(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`BigDataPoolsOperations` - * 2021-06-01-preview: :class:`BigDataPoolsOperations` - """ - api_version = self._get_api_version('big_data_pools') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import BigDataPoolsOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import BigDataPoolsOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'big_data_pools'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def data_flow(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`DataFlowOperations` - * 2021-06-01-preview: :class:`DataFlowOperations` - """ - api_version = self._get_api_version('data_flow') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import DataFlowOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import DataFlowOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'data_flow'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def data_flow_debug_session(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`DataFlowDebugSessionOperations` - * 2021-06-01-preview: :class:`DataFlowDebugSessionOperations` - """ - api_version = self._get_api_version('data_flow_debug_session') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import DataFlowDebugSessionOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import DataFlowDebugSessionOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'data_flow_debug_session'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def dataset(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`DatasetOperations` - * 2021-06-01-preview: :class:`DatasetOperations` - """ - api_version = self._get_api_version('dataset') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import DatasetOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import DatasetOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'dataset'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def integration_runtimes(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`IntegrationRuntimesOperations` - * 2021-06-01-preview: :class:`IntegrationRuntimesOperations` - """ - api_version = self._get_api_version('integration_runtimes') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import IntegrationRuntimesOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import IntegrationRuntimesOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'integration_runtimes'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def library(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`LibraryOperations` - * 2021-06-01-preview: :class:`LibraryOperations` - """ - api_version = self._get_api_version('library') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import LibraryOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import LibraryOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'library'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def linked_service(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`LinkedServiceOperations` - * 2021-06-01-preview: :class:`LinkedServiceOperations` - """ - api_version = self._get_api_version('linked_service') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import LinkedServiceOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import LinkedServiceOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'linked_service'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def notebook(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`NotebookOperations` - * 2021-06-01-preview: :class:`NotebookOperations` - """ - api_version = self._get_api_version('notebook') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import NotebookOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import NotebookOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'notebook'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def notebook_operation_result(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`NotebookOperationResultOperations` - * 2021-06-01-preview: :class:`NotebookOperationResultOperations` - """ - api_version = self._get_api_version('notebook_operation_result') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import NotebookOperationResultOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import NotebookOperationResultOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'notebook_operation_result'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - @property - def operation_result(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`OperationResultOperations` - * 2021-06-01-preview: :class:`OperationResultOperations` - """ - api_version = self._get_api_version('operation_result') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import OperationResultOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import OperationResultOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'operation_result'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def operation_status(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`OperationStatusOperations` - * 2021-06-01-preview: :class:`OperationStatusOperations` - """ - api_version = self._get_api_version('operation_status') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import OperationStatusOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import OperationStatusOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'operation_status'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def pipeline(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`PipelineOperations` - * 2021-06-01-preview: :class:`PipelineOperations` - """ - api_version = self._get_api_version('pipeline') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import PipelineOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import PipelineOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'pipeline'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def pipeline_run(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`PipelineRunOperations` - * 2021-06-01-preview: :class:`PipelineRunOperations` - """ - api_version = self._get_api_version('pipeline_run') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import PipelineRunOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import PipelineRunOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'pipeline_run'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def spark_configuration(self): - """Instance depends on the API version: - - * 2021-06-01-preview: :class:`SparkConfigurationOperations` - """ - api_version = self._get_api_version('spark_configuration') - if api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import SparkConfigurationOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'spark_configuration'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def spark_job_definition(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`SparkJobDefinitionOperations` - * 2021-06-01-preview: :class:`SparkJobDefinitionOperations` - """ - api_version = self._get_api_version('spark_job_definition') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import SparkJobDefinitionOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import SparkJobDefinitionOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'spark_job_definition'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def sql_pools(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`SqlPoolsOperations` - * 2021-06-01-preview: :class:`SqlPoolsOperations` - """ - api_version = self._get_api_version('sql_pools') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import SqlPoolsOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import SqlPoolsOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'sql_pools'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def sql_script(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`SqlScriptOperations` - * 2021-06-01-preview: :class:`SqlScriptOperations` - """ - api_version = self._get_api_version('sql_script') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import SqlScriptOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import SqlScriptOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'sql_script'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def trigger(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`TriggerOperations` - * 2021-06-01-preview: :class:`TriggerOperations` - """ - api_version = self._get_api_version('trigger') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import TriggerOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import TriggerOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'trigger'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def trigger_run(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`TriggerRunOperations` - * 2021-06-01-preview: :class:`TriggerRunOperations` - """ - api_version = self._get_api_version('trigger_run') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import TriggerRunOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import TriggerRunOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'trigger_run'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def workspace(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`WorkspaceOperations` - * 2021-06-01-preview: :class:`WorkspaceOperations` - """ - api_version = self._get_api_version('workspace') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import WorkspaceOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import WorkspaceOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'workspace'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - @property - def workspace_git_repo_management(self): - """Instance depends on the API version: - - * 2020-12-01: :class:`WorkspaceGitRepoManagementOperations` - * 2021-06-01-preview: :class:`WorkspaceGitRepoManagementOperations` - """ - api_version = self._get_api_version('workspace_git_repo_management') - if api_version == '2020-12-01': - from ..v2020_12_01.aio.operations import WorkspaceGitRepoManagementOperations as OperationClass - elif api_version == '2021-06-01-preview': - from ..v2021_06_01_preview.aio.operations import WorkspaceGitRepoManagementOperations as OperationClass - else: - raise ValueError("API version {} does not have operation group 'workspace_git_repo_management'".format(api_version)) - return OperationClass(self._client, self._config, Serializer(self._models_dict(api_version)), Deserializer(self._models_dict(api_version))) - - async def close(self): + client_models = {k: v for k, v in models.__dict__.items() if isinstance(v, type)} + self._serialize = Serializer(client_models) + self._serialize.client_side_validation = False + self._deserialize = Deserializer(client_models) + + self.kql_scripts = KqlScriptsOperations( + self._client, self._config, self._serialize, self._deserialize) + self.kql_script = KqlScriptOperations( + self._client, self._config, self._serialize, self._deserialize) + self.spark_configuration = SparkConfigurationOperations( + self._client, self._config, self._serialize, self._deserialize) + self.big_data_pools = BigDataPoolsOperations( + self._client, self._config, self._serialize, self._deserialize) + self.data_flow = DataFlowOperations( + self._client, self._config, self._serialize, self._deserialize) + self.data_flow_debug_session = DataFlowDebugSessionOperations( + self._client, self._config, self._serialize, self._deserialize) + self.dataset = DatasetOperations( + self._client, self._config, self._serialize, self._deserialize) + self.workspace_git_repo_management = WorkspaceGitRepoManagementOperations( + self._client, self._config, self._serialize, self._deserialize) + self.integration_runtimes = IntegrationRuntimesOperations( + self._client, self._config, self._serialize, self._deserialize) + self.library = LibraryOperations( + self._client, self._config, self._serialize, self._deserialize) + self.linked_service = LinkedServiceOperations( + self._client, self._config, self._serialize, self._deserialize) + self.notebook = NotebookOperations( + self._client, self._config, self._serialize, self._deserialize) + self.notebook_operation_result = NotebookOperationResultOperations( + self._client, self._config, self._serialize, self._deserialize) + self.pipeline = PipelineOperations( + self._client, self._config, self._serialize, self._deserialize) + self.pipeline_run = PipelineRunOperations( + self._client, self._config, self._serialize, self._deserialize) + self.spark_job_definition = SparkJobDefinitionOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_pools = SqlPoolsOperations( + self._client, self._config, self._serialize, self._deserialize) + self.sql_script = SqlScriptOperations( + self._client, self._config, self._serialize, self._deserialize) + self.trigger = TriggerOperations( + self._client, self._config, self._serialize, self._deserialize) + self.trigger_run = TriggerRunOperations( + self._client, self._config, self._serialize, self._deserialize) + self.workspace = WorkspaceOperations( + self._client, self._config, self._serialize, self._deserialize) + + async def _send_request(self, http_request: HttpRequest, **kwargs: Any) -> AsyncHttpResponse: + """Runs the network request through the client's chained policies. + + :param http_request: The network request you want to make. Required. + :type http_request: ~azure.core.pipeline.transport.HttpRequest + :keyword bool stream: Whether the response payload will be streamed. Defaults to True. + :return: The response of your network call. Does not do error handling on your response. + :rtype: ~azure.core.pipeline.transport.AsyncHttpResponse + """ + path_format_arguments = { + 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), + } + http_request.url = self._client.format_url(http_request.url, **path_format_arguments) + stream = kwargs.pop("stream", True) + pipeline_response = await self._client._pipeline.run(http_request, stream=stream, **kwargs) + return pipeline_response.http_response + + async def close(self) -> None: await self._client.close() - async def __aenter__(self): + + async def __aenter__(self) -> "ArtifactsClient": await self._client.__aenter__() return self - async def __aexit__(self, *exc_details): + + async def __aexit__(self, *exc_details) -> None: await self._client.__aexit__(*exc_details) diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/_configuration.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/_configuration.py index 2025d8f6f9fd..0d67f9417b72 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/_configuration.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/_configuration.py @@ -1,13 +1,11 @@ # coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# +# Licensed under the MIT License. See License.txt in the project root for license information. # Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- + from typing import Any, TYPE_CHECKING from azure.core.configuration import Configuration @@ -19,6 +17,7 @@ # pylint: disable=unused-import,ungrouped-imports from azure.core.credentials_async import AsyncTokenCredential + class ArtifactsClientConfiguration(Configuration): """Configuration for ArtifactsClient. @@ -35,7 +34,7 @@ def __init__( self, credential: "AsyncTokenCredential", endpoint: str, - **kwargs # type: Any + **kwargs: Any ) -> None: if credential is None: raise ValueError("Parameter 'credential' must not be None.") @@ -46,7 +45,7 @@ def __init__( self.credential = credential self.endpoint = endpoint self.credential_scopes = kwargs.pop('credential_scopes', ['https://dev.azuresynapse.net/.default']) - kwargs.setdefault('sdk_moniker', 'azure-synapse-artifacts/{}'.format(VERSION)) + kwargs.setdefault('sdk_moniker', 'synapse-artifacts/{}'.format(VERSION)) self._configure(**kwargs) def _configure( diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/__init__.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/__init__.py similarity index 91% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/__init__.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/__init__.py index 4620c478f14b..7b2bd054134a 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/__init__.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/__init__.py @@ -6,48 +6,48 @@ # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from ._linked_service_operations import LinkedServiceOperations -from ._dataset_operations import DatasetOperations -from ._pipeline_operations import PipelineOperations -from ._pipeline_run_operations import PipelineRunOperations -from ._trigger_operations import TriggerOperations -from ._trigger_run_operations import TriggerRunOperations -from ._data_flow_operations import DataFlowOperations -from ._data_flow_debug_session_operations import DataFlowDebugSessionOperations -from ._sql_script_operations import SqlScriptOperations -from ._spark_job_definition_operations import SparkJobDefinitionOperations -from ._notebook_operations import NotebookOperations -from ._notebook_operation_result_operations import NotebookOperationResultOperations +from ._kql_scripts_operations import KqlScriptsOperations +from ._kql_script_operations import KqlScriptOperations from ._spark_configuration_operations import SparkConfigurationOperations from ._big_data_pools_operations import BigDataPoolsOperations +from ._data_flow_operations import DataFlowOperations +from ._data_flow_debug_session_operations import DataFlowDebugSessionOperations +from ._dataset_operations import DatasetOperations from ._workspace_git_repo_management_operations import WorkspaceGitRepoManagementOperations from ._integration_runtimes_operations import IntegrationRuntimesOperations from ._library_operations import LibraryOperations -from ._operation_result_operations import OperationResultOperations -from ._operation_status_operations import OperationStatusOperations +from ._linked_service_operations import LinkedServiceOperations +from ._notebook_operations import NotebookOperations +from ._notebook_operation_result_operations import NotebookOperationResultOperations +from ._pipeline_operations import PipelineOperations +from ._pipeline_run_operations import PipelineRunOperations +from ._spark_job_definition_operations import SparkJobDefinitionOperations from ._sql_pools_operations import SqlPoolsOperations +from ._sql_script_operations import SqlScriptOperations +from ._trigger_operations import TriggerOperations +from ._trigger_run_operations import TriggerRunOperations from ._workspace_operations import WorkspaceOperations __all__ = [ - 'LinkedServiceOperations', - 'DatasetOperations', - 'PipelineOperations', - 'PipelineRunOperations', - 'TriggerOperations', - 'TriggerRunOperations', - 'DataFlowOperations', - 'DataFlowDebugSessionOperations', - 'SqlScriptOperations', - 'SparkJobDefinitionOperations', - 'NotebookOperations', - 'NotebookOperationResultOperations', + 'KqlScriptsOperations', + 'KqlScriptOperations', 'SparkConfigurationOperations', 'BigDataPoolsOperations', + 'DataFlowOperations', + 'DataFlowDebugSessionOperations', + 'DatasetOperations', 'WorkspaceGitRepoManagementOperations', 'IntegrationRuntimesOperations', 'LibraryOperations', - 'OperationResultOperations', - 'OperationStatusOperations', + 'LinkedServiceOperations', + 'NotebookOperations', + 'NotebookOperationResultOperations', + 'PipelineOperations', + 'PipelineRunOperations', + 'SparkJobDefinitionOperations', 'SqlPoolsOperations', + 'SqlScriptOperations', + 'TriggerOperations', + 'TriggerRunOperations', 'WorkspaceOperations', ] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_big_data_pools_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_big_data_pools_operations.py similarity index 96% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_big_data_pools_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_big_data_pools_operations.py index 577710f0d995..7da9400cfad0 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_big_data_pools_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_big_data_pools_operations.py @@ -24,7 +24,7 @@ class BigDataPoolsOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -47,7 +47,7 @@ async def list( :keyword callable cls: A custom type or function that will be passed the direct response :return: BigDataPoolResourceInfoListResult, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.BigDataPoolResourceInfoListResult + :rtype: ~azure.synapse.artifacts.models.BigDataPoolResourceInfoListResult :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.BigDataPoolResourceInfoListResult"] @@ -101,7 +101,7 @@ async def get( :type big_data_pool_name: str :keyword callable cls: A custom type or function that will be passed the direct response :return: BigDataPoolResourceInfo, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.BigDataPoolResourceInfo + :rtype: ~azure.synapse.artifacts.models.BigDataPoolResourceInfo :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.BigDataPoolResourceInfo"] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_data_flow_debug_session_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_data_flow_debug_session_operations.py similarity index 96% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_data_flow_debug_session_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_data_flow_debug_session_operations.py index 6eccf6121e82..3eb0d3913a6e 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_data_flow_debug_session_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_data_flow_debug_session_operations.py @@ -27,7 +27,7 @@ class DataFlowDebugSessionOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -81,7 +81,7 @@ async def _create_data_flow_debug_session_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) response_headers = {} @@ -106,7 +106,7 @@ async def begin_create_data_flow_debug_session( """Creates a data flow debug session. :param request: Data flow debug session definition. - :type request: ~azure.synapse.artifacts.v2020_12_01.models.CreateDataFlowDebugSessionRequest + :type request: ~azure.synapse.artifacts.models.CreateDataFlowDebugSessionRequest :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: By default, your polling method will be AsyncLROBasePolling. @@ -114,7 +114,7 @@ async def begin_create_data_flow_debug_session( :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of AsyncLROPoller that returns either CreateDataFlowDebugSessionResponse or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2020_12_01.models.CreateDataFlowDebugSessionResponse] + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.models.CreateDataFlowDebugSessionResponse] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] @@ -167,7 +167,7 @@ def query_data_flow_debug_sessions_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either QueryDataFlowDebugSessionsResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2020_12_01.models.QueryDataFlowDebugSessionsResponse] + :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.models.QueryDataFlowDebugSessionsResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.QueryDataFlowDebugSessionsResponse"] @@ -219,7 +219,7 @@ async def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -238,10 +238,10 @@ async def add_data_flow( """Add a data flow into debug session. :param request: Data flow debug session definition with debug content. - :type request: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowDebugPackage + :type request: ~azure.synapse.artifacts.models.DataFlowDebugPackage :keyword callable cls: A custom type or function that will be passed the direct response :return: AddDataFlowToDebugSessionResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.AddDataFlowToDebugSessionResponse + :rtype: ~azure.synapse.artifacts.models.AddDataFlowToDebugSessionResponse :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.AddDataFlowToDebugSessionResponse"] @@ -278,7 +278,7 @@ async def add_data_flow( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = self._deserialize('AddDataFlowToDebugSessionResponse', pipeline_response) @@ -297,7 +297,7 @@ async def delete_data_flow_debug_session( """Deletes a data flow debug session. :param request: Data flow debug session definition for deletion. - :type request: ~azure.synapse.artifacts.v2020_12_01.models.DeleteDataFlowDebugSessionRequest + :type request: ~azure.synapse.artifacts.models.DeleteDataFlowDebugSessionRequest :keyword callable cls: A custom type or function that will be passed the direct response :return: None, or the result of cls(response) :rtype: None @@ -337,7 +337,7 @@ async def delete_data_flow_debug_session( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -384,7 +384,7 @@ async def _execute_command_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) response_headers = {} @@ -409,7 +409,7 @@ async def begin_execute_command( """Execute a data flow debug command. :param request: Data flow debug command definition. - :type request: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowDebugCommandRequest + :type request: ~azure.synapse.artifacts.models.DataFlowDebugCommandRequest :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: By default, your polling method will be AsyncLROBasePolling. @@ -417,7 +417,7 @@ async def begin_execute_command( :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of AsyncLROPoller that returns either DataFlowDebugCommandResponse or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2020_12_01.models.DataFlowDebugCommandResponse] + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.models.DataFlowDebugCommandResponse] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_data_flow_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_data_flow_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_data_flow_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_data_flow_operations.py index 85a9f079b00a..acc23d22d50e 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_data_flow_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_data_flow_operations.py @@ -27,7 +27,7 @@ class DataFlowOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -88,7 +88,7 @@ async def _create_or_update_data_flow_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -113,7 +113,7 @@ async def begin_create_or_update_data_flow( :param data_flow_name: The data flow name. :type data_flow_name: str :param properties: Data flow properties. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.DataFlow + :type properties: ~azure.synapse.artifacts.models.DataFlow :param if_match: ETag of the data flow entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -124,7 +124,7 @@ async def begin_create_or_update_data_flow( :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of AsyncLROPoller that returns either DataFlowResource or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2020_12_01.models.DataFlowResource] + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.models.DataFlowResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] @@ -187,7 +187,7 @@ async def get_data_flow( :type if_none_match: str :keyword callable cls: A custom type or function that will be passed the direct response :return: DataFlowResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowResource + :rtype: ~azure.synapse.artifacts.models.DataFlowResource :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.DataFlowResource"] @@ -222,7 +222,7 @@ async def get_data_flow( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = self._deserialize('DataFlowResource', pipeline_response) @@ -268,7 +268,7 @@ async def _delete_data_flow_initial( if response.status_code not in [200, 202, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -378,7 +378,7 @@ async def _rename_data_flow_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -457,7 +457,7 @@ def get_data_flows_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either DataFlowListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2020_12_01.models.DataFlowListResponse] + :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.models.DataFlowListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.DataFlowListResponse"] @@ -509,7 +509,7 @@ async def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_dataset_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_dataset_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_dataset_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_dataset_operations.py index b35d8aa3a904..a5fd29f27deb 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_dataset_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_dataset_operations.py @@ -27,7 +27,7 @@ class DatasetOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -50,7 +50,7 @@ def get_datasets_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either DatasetListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2020_12_01.models.DatasetListResponse] + :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.models.DatasetListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.DatasetListResponse"] @@ -102,7 +102,7 @@ async def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -159,7 +159,7 @@ async def _create_or_update_dataset_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -184,7 +184,7 @@ async def begin_create_or_update_dataset( :param dataset_name: The dataset name. :type dataset_name: str :param properties: Dataset properties. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.Dataset + :type properties: ~azure.synapse.artifacts.models.Dataset :param if_match: ETag of the dataset entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -195,7 +195,7 @@ async def begin_create_or_update_dataset( :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of AsyncLROPoller that returns either DatasetResource or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2020_12_01.models.DatasetResource] + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.models.DatasetResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] @@ -258,7 +258,7 @@ async def get_dataset( :type if_none_match: str :keyword callable cls: A custom type or function that will be passed the direct response :return: DatasetResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.DatasetResource or None + :rtype: ~azure.synapse.artifacts.models.DatasetResource or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.DatasetResource"]] @@ -293,7 +293,7 @@ async def get_dataset( if response.status_code not in [200, 304]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -341,7 +341,7 @@ async def _delete_dataset_initial( if response.status_code not in [200, 202, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -451,7 +451,7 @@ async def _rename_dataset_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_integration_runtimes_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_integration_runtimes_operations.py similarity index 96% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_integration_runtimes_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_integration_runtimes_operations.py index 0175fea49ade..37bc7449f5b9 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_integration_runtimes_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_integration_runtimes_operations.py @@ -24,7 +24,7 @@ class IntegrationRuntimesOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -47,7 +47,7 @@ async def list( :keyword callable cls: A custom type or function that will be passed the direct response :return: IntegrationRuntimeListResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeListResponse + :rtype: ~azure.synapse.artifacts.models.IntegrationRuntimeListResponse :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.IntegrationRuntimeListResponse"] @@ -101,7 +101,7 @@ async def get( :type integration_runtime_name: str :keyword callable cls: A custom type or function that will be passed the direct response :return: IntegrationRuntimeResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeResource + :rtype: ~azure.synapse.artifacts.models.IntegrationRuntimeResource :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.IntegrationRuntimeResource"] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_kql_script_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_kql_script_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_kql_script_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_kql_script_operations.py index 03d1e09d3a5b..94a5dfe6c420 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_kql_script_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_kql_script_operations.py @@ -26,7 +26,7 @@ class KqlScriptOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -106,7 +106,7 @@ async def begin_create_or_update( :param kql_script_name: KQL script name. :type kql_script_name: str :param kql_script: KQL script. - :type kql_script: ~azure.synapse.artifacts.v2021_06_01_preview.models.KqlScriptResource + :type kql_script: ~azure.synapse.artifacts.models.KqlScriptResource :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: By default, your polling method will be AsyncLROBasePolling. @@ -114,7 +114,7 @@ async def begin_create_or_update( :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of AsyncLROPoller that returns either KqlScriptResource or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.KqlScriptResource] + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.models.KqlScriptResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] @@ -172,7 +172,7 @@ async def get_by_name( :type kql_script_name: str :keyword callable cls: A custom type or function that will be passed the direct response :return: KqlScriptResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.KqlScriptResource + :rtype: ~azure.synapse.artifacts.models.KqlScriptResource :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.KqlScriptResource"] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_kql_scripts_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_kql_scripts_operations.py similarity index 97% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_kql_scripts_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_kql_scripts_operations.py index fb3bf363ab00..eda4e14ce167 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_kql_scripts_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_kql_scripts_operations.py @@ -25,7 +25,7 @@ class KqlScriptsOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -48,7 +48,7 @@ def get_all( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either KqlScriptsResourceCollectionResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.KqlScriptsResourceCollectionResponse] + :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.models.KqlScriptsResourceCollectionResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.KqlScriptsResourceCollectionResponse"] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_library_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_library_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_library_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_library_operations.py index ff3f9bc1f63a..b39d26195e1c 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_library_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_library_operations.py @@ -27,7 +27,7 @@ class LibraryOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -50,7 +50,7 @@ def list( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either LibraryListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2020_12_01.models.LibraryListResponse] + :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.models.LibraryListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.LibraryListResponse"] @@ -102,7 +102,7 @@ async def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -148,7 +148,7 @@ async def _flush_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -235,7 +235,7 @@ async def get_operation_result( :type operation_id: str :keyword callable cls: A custom type or function that will be passed the direct response :return: LibraryResource or OperationResult, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.LibraryResource or ~azure.synapse.artifacts.v2020_12_01.models.OperationResult + :rtype: ~azure.synapse.artifacts.models.LibraryResource or ~azure.synapse.artifacts.models.OperationResult :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Union["_models.LibraryResource", "_models.OperationResult"]] @@ -268,7 +268,7 @@ async def get_operation_result( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if response.status_code == 200: @@ -318,7 +318,7 @@ async def _delete_initial( if response.status_code not in [200, 202, 409]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -406,7 +406,7 @@ async def get( :type library_name: str :keyword callable cls: A custom type or function that will be passed the direct response :return: LibraryResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.LibraryResource or None + :rtype: ~azure.synapse.artifacts.models.LibraryResource or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.LibraryResource"]] @@ -439,7 +439,7 @@ async def get( if response.status_code not in [200, 304]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -487,7 +487,7 @@ async def _create_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -626,7 +626,7 @@ async def append( if response.status_code not in [201]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_linked_service_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_linked_service_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_linked_service_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_linked_service_operations.py index 01f3c4c8b9ef..bbf6b3edb226 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_linked_service_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_linked_service_operations.py @@ -27,7 +27,7 @@ class LinkedServiceOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -50,7 +50,7 @@ def get_linked_services_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either LinkedServiceListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceListResponse] + :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.models.LinkedServiceListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.LinkedServiceListResponse"] @@ -102,7 +102,7 @@ async def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -159,7 +159,7 @@ async def _create_or_update_linked_service_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -184,7 +184,7 @@ async def begin_create_or_update_linked_service( :param linked_service_name: The linked service name. :type linked_service_name: str :param properties: Properties of linked service. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.LinkedService + :type properties: ~azure.synapse.artifacts.models.LinkedService :param if_match: ETag of the linkedService entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -195,7 +195,7 @@ async def begin_create_or_update_linked_service( :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of AsyncLROPoller that returns either LinkedServiceResource or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceResource] + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.models.LinkedServiceResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] @@ -259,7 +259,7 @@ async def get_linked_service( :type if_none_match: str :keyword callable cls: A custom type or function that will be passed the direct response :return: LinkedServiceResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceResource or None + :rtype: ~azure.synapse.artifacts.models.LinkedServiceResource or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.LinkedServiceResource"]] @@ -294,7 +294,7 @@ async def get_linked_service( if response.status_code not in [200, 304]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -342,7 +342,7 @@ async def _delete_linked_service_initial( if response.status_code not in [200, 202, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -452,7 +452,7 @@ async def _rename_linked_service_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_notebook_operation_result_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_notebook_operation_result_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_notebook_operation_result_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_notebook_operation_result_operations.py index 8c058583a9b0..c82448dd85e7 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_notebook_operation_result_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_notebook_operation_result_operations.py @@ -24,7 +24,7 @@ class NotebookOperationResultOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_notebook_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_notebook_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_notebook_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_notebook_operations.py index d9b380cc1f08..f918094d56cd 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_notebook_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_notebook_operations.py @@ -27,7 +27,7 @@ class NotebookOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -50,7 +50,7 @@ def get_notebooks_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either NotebookListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2020_12_01.models.NotebookListResponse] + :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.models.NotebookListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.NotebookListResponse"] @@ -102,7 +102,7 @@ async def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -121,7 +121,7 @@ def get_notebook_summary_by_work_space( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either NotebookListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2020_12_01.models.NotebookListResponse] + :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.models.NotebookListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.NotebookListResponse"] @@ -173,7 +173,7 @@ async def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -228,7 +228,7 @@ async def _create_or_update_notebook_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -253,7 +253,7 @@ async def begin_create_or_update_notebook( :param notebook_name: The notebook name. :type notebook_name: str :param notebook: Note book resource definition. - :type notebook: ~azure.synapse.artifacts.v2020_12_01.models.NotebookResource + :type notebook: ~azure.synapse.artifacts.models.NotebookResource :param if_match: ETag of the Note book entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -264,7 +264,7 @@ async def begin_create_or_update_notebook( :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of AsyncLROPoller that returns either NotebookResource or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2020_12_01.models.NotebookResource] + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.models.NotebookResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] @@ -327,7 +327,7 @@ async def get_notebook( :type if_none_match: str :keyword callable cls: A custom type or function that will be passed the direct response :return: NotebookResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.NotebookResource or None + :rtype: ~azure.synapse.artifacts.models.NotebookResource or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.NotebookResource"]] @@ -362,7 +362,7 @@ async def get_notebook( if response.status_code not in [200, 304]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -410,7 +410,7 @@ async def _delete_notebook_initial( if response.status_code not in [200, 202, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -520,7 +520,7 @@ async def _rename_notebook_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_pipeline_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_pipeline_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_pipeline_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_pipeline_operations.py index f5a8991e9b7f..2da105e2efb8 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_pipeline_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_pipeline_operations.py @@ -27,7 +27,7 @@ class PipelineOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -50,7 +50,7 @@ def get_pipelines_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either PipelineListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2020_12_01.models.PipelineListResponse] + :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.models.PipelineListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.PipelineListResponse"] @@ -102,7 +102,7 @@ async def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -157,7 +157,7 @@ async def _create_or_update_pipeline_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -182,7 +182,7 @@ async def begin_create_or_update_pipeline( :param pipeline_name: The pipeline name. :type pipeline_name: str :param pipeline: Pipeline resource definition. - :type pipeline: ~azure.synapse.artifacts.v2020_12_01.models.PipelineResource + :type pipeline: ~azure.synapse.artifacts.models.PipelineResource :param if_match: ETag of the pipeline entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -193,7 +193,7 @@ async def begin_create_or_update_pipeline( :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of AsyncLROPoller that returns either PipelineResource or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2020_12_01.models.PipelineResource] + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.models.PipelineResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] @@ -256,7 +256,7 @@ async def get_pipeline( :type if_none_match: str :keyword callable cls: A custom type or function that will be passed the direct response :return: PipelineResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.PipelineResource or None + :rtype: ~azure.synapse.artifacts.models.PipelineResource or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.PipelineResource"]] @@ -291,7 +291,7 @@ async def get_pipeline( if response.status_code not in [200, 304]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -339,7 +339,7 @@ async def _delete_pipeline_initial( if response.status_code not in [200, 202, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -449,7 +449,7 @@ async def _rename_pipeline_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -547,7 +547,7 @@ async def create_pipeline_run( :type parameters: dict[str, any] :keyword callable cls: A custom type or function that will be passed the direct response :return: CreateRunResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.CreateRunResponse + :rtype: ~azure.synapse.artifacts.models.CreateRunResponse :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.CreateRunResponse"] @@ -594,7 +594,7 @@ async def create_pipeline_run( if response.status_code not in [202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = self._deserialize('CreateRunResponse', pipeline_response) diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_pipeline_run_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_pipeline_run_operations.py similarity index 95% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_pipeline_run_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_pipeline_run_operations.py index b3d8c577454f..c19ccfdb10ae 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_pipeline_run_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_pipeline_run_operations.py @@ -24,7 +24,7 @@ class PipelineRunOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -47,10 +47,10 @@ async def query_pipeline_runs_by_workspace( """Query pipeline runs in the workspace based on input filter conditions. :param filter_parameters: Parameters to filter the pipeline run. - :type filter_parameters: ~azure.synapse.artifacts.v2020_12_01.models.RunFilterParameters + :type filter_parameters: ~azure.synapse.artifacts.models.RunFilterParameters :keyword callable cls: A custom type or function that will be passed the direct response :return: PipelineRunsQueryResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.PipelineRunsQueryResponse + :rtype: ~azure.synapse.artifacts.models.PipelineRunsQueryResponse :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.PipelineRunsQueryResponse"] @@ -87,7 +87,7 @@ async def query_pipeline_runs_by_workspace( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = self._deserialize('PipelineRunsQueryResponse', pipeline_response) @@ -109,7 +109,7 @@ async def get_pipeline_run( :type run_id: str :keyword callable cls: A custom type or function that will be passed the direct response :return: PipelineRun, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.PipelineRun + :rtype: ~azure.synapse.artifacts.models.PipelineRun :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.PipelineRun"] @@ -142,7 +142,7 @@ async def get_pipeline_run( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = self._deserialize('PipelineRun', pipeline_response) @@ -167,10 +167,10 @@ async def query_activity_runs( :param run_id: The pipeline run identifier. :type run_id: str :param filter_parameters: Parameters to filter the activity runs. - :type filter_parameters: ~azure.synapse.artifacts.v2020_12_01.models.RunFilterParameters + :type filter_parameters: ~azure.synapse.artifacts.models.RunFilterParameters :keyword callable cls: A custom type or function that will be passed the direct response :return: ActivityRunsQueryResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.ActivityRunsQueryResponse + :rtype: ~azure.synapse.artifacts.models.ActivityRunsQueryResponse :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.ActivityRunsQueryResponse"] @@ -209,7 +209,7 @@ async def query_activity_runs( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = self._deserialize('ActivityRunsQueryResponse', pipeline_response) @@ -270,7 +270,7 @@ async def cancel_pipeline_run( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_spark_configuration_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_spark_configuration_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_spark_configuration_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_spark_configuration_operations.py index ecb99ccf0b2b..9c8c686b0e3e 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_spark_configuration_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_spark_configuration_operations.py @@ -27,7 +27,7 @@ class SparkConfigurationOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -50,7 +50,7 @@ def get_spark_configurations_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either SparkConfigurationListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.SparkConfigurationListResponse] + :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.models.SparkConfigurationListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.SparkConfigurationListResponse"] @@ -184,7 +184,7 @@ async def begin_create_or_update_spark_configuration( :param spark_configuration_name: The spark Configuration name. :type spark_configuration_name: str :param properties: Properties of Spark Configuration. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkConfiguration + :type properties: ~azure.synapse.artifacts.models.SparkConfiguration :param if_match: ETag of the sparkConfiguration entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -195,7 +195,7 @@ async def begin_create_or_update_spark_configuration( :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of AsyncLROPoller that returns either SparkConfigurationResource or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.SparkConfigurationResource] + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.models.SparkConfigurationResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] @@ -259,7 +259,7 @@ async def get_spark_configuration( :type if_none_match: str :keyword callable cls: A custom type or function that will be passed the direct response :return: SparkConfigurationResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkConfigurationResource or None + :rtype: ~azure.synapse.artifacts.models.SparkConfigurationResource or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.SparkConfigurationResource"]] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_spark_job_definition_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_spark_job_definition_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_spark_job_definition_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_spark_job_definition_operations.py index bbf2b5b77e1c..1c7787e99d0c 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_spark_job_definition_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_spark_job_definition_operations.py @@ -27,7 +27,7 @@ class SparkJobDefinitionOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -50,7 +50,7 @@ def get_spark_job_definitions_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either SparkJobDefinitionsListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2020_12_01.models.SparkJobDefinitionsListResponse] + :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.models.SparkJobDefinitionsListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.SparkJobDefinitionsListResponse"] @@ -102,7 +102,7 @@ async def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -159,7 +159,7 @@ async def _create_or_update_spark_job_definition_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -184,7 +184,7 @@ async def begin_create_or_update_spark_job_definition( :param spark_job_definition_name: The spark job definition name. :type spark_job_definition_name: str :param properties: Properties of spark job definition. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.SparkJobDefinition + :type properties: ~azure.synapse.artifacts.models.SparkJobDefinition :param if_match: ETag of the Spark Job Definition entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -195,7 +195,7 @@ async def begin_create_or_update_spark_job_definition( :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of AsyncLROPoller that returns either SparkJobDefinitionResource or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2020_12_01.models.SparkJobDefinitionResource] + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.models.SparkJobDefinitionResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] @@ -259,7 +259,7 @@ async def get_spark_job_definition( :type if_none_match: str :keyword callable cls: A custom type or function that will be passed the direct response :return: SparkJobDefinitionResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.SparkJobDefinitionResource or None + :rtype: ~azure.synapse.artifacts.models.SparkJobDefinitionResource or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.SparkJobDefinitionResource"]] @@ -294,7 +294,7 @@ async def get_spark_job_definition( if response.status_code not in [200, 304]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -342,7 +342,7 @@ async def _delete_spark_job_definition_initial( if response.status_code not in [200, 202, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -444,7 +444,7 @@ async def _execute_spark_job_definition_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if response.status_code == 200: @@ -475,7 +475,7 @@ async def begin_execute_spark_job_definition( :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of AsyncLROPoller that returns either SparkBatchJob or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2020_12_01.models.SparkBatchJob] + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.models.SparkBatchJob] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] @@ -564,7 +564,7 @@ async def _rename_spark_job_definition_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -676,7 +676,7 @@ async def _debug_spark_job_definition_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if response.status_code == 200: @@ -699,7 +699,7 @@ async def begin_debug_spark_job_definition( """Debug the spark job definition. :param properties: Properties of spark job definition. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.SparkJobDefinition + :type properties: ~azure.synapse.artifacts.models.SparkJobDefinition :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: By default, your polling method will be AsyncLROBasePolling. @@ -707,7 +707,7 @@ async def begin_debug_spark_job_definition( :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of AsyncLROPoller that returns either SparkBatchJob or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2020_12_01.models.SparkBatchJob] + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.models.SparkBatchJob] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_sql_pools_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_sql_pools_operations.py similarity index 96% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_sql_pools_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_sql_pools_operations.py index 4b565d881623..e1a0f5dbf6ad 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_sql_pools_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_sql_pools_operations.py @@ -24,7 +24,7 @@ class SqlPoolsOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -47,7 +47,7 @@ async def list( :keyword callable cls: A custom type or function that will be passed the direct response :return: SqlPoolInfoListResult, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.SqlPoolInfoListResult + :rtype: ~azure.synapse.artifacts.models.SqlPoolInfoListResult :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.SqlPoolInfoListResult"] @@ -101,7 +101,7 @@ async def get( :type sql_pool_name: str :keyword callable cls: A custom type or function that will be passed the direct response :return: SqlPool, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.SqlPool + :rtype: ~azure.synapse.artifacts.models.SqlPool :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.SqlPool"] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_sql_script_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_sql_script_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_sql_script_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_sql_script_operations.py index 6875b17dcb41..511cef904e3e 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_sql_script_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_sql_script_operations.py @@ -27,7 +27,7 @@ class SqlScriptOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -50,7 +50,7 @@ def get_sql_scripts_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either SqlScriptsListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2020_12_01.models.SqlScriptsListResponse] + :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.models.SqlScriptsListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.SqlScriptsListResponse"] @@ -102,7 +102,7 @@ async def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -157,7 +157,7 @@ async def _create_or_update_sql_script_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -182,7 +182,7 @@ async def begin_create_or_update_sql_script( :param sql_script_name: The sql script name. :type sql_script_name: str :param sql_script: Sql Script resource definition. - :type sql_script: ~azure.synapse.artifacts.v2020_12_01.models.SqlScriptResource + :type sql_script: ~azure.synapse.artifacts.models.SqlScriptResource :param if_match: ETag of the SQL script entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -193,7 +193,7 @@ async def begin_create_or_update_sql_script( :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of AsyncLROPoller that returns either SqlScriptResource or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2020_12_01.models.SqlScriptResource] + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.models.SqlScriptResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] @@ -256,7 +256,7 @@ async def get_sql_script( :type if_none_match: str :keyword callable cls: A custom type or function that will be passed the direct response :return: SqlScriptResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.SqlScriptResource or None + :rtype: ~azure.synapse.artifacts.models.SqlScriptResource or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.SqlScriptResource"]] @@ -291,7 +291,7 @@ async def get_sql_script( if response.status_code not in [200, 304]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -339,7 +339,7 @@ async def _delete_sql_script_initial( if response.status_code not in [200, 202, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -449,7 +449,7 @@ async def _rename_sql_script_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_trigger_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_trigger_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_trigger_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_trigger_operations.py index 2eeb015a5f89..3c39915184a7 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_trigger_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_trigger_operations.py @@ -27,7 +27,7 @@ class TriggerOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -50,7 +50,7 @@ def get_triggers_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either TriggerListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2020_12_01.models.TriggerListResponse] + :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.models.TriggerListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.TriggerListResponse"] @@ -102,7 +102,7 @@ async def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -159,7 +159,7 @@ async def _create_or_update_trigger_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -184,7 +184,7 @@ async def begin_create_or_update_trigger( :param trigger_name: The trigger name. :type trigger_name: str :param properties: Properties of the trigger. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.Trigger + :type properties: ~azure.synapse.artifacts.models.Trigger :param if_match: ETag of the trigger entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -195,7 +195,7 @@ async def begin_create_or_update_trigger( :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of AsyncLROPoller that returns either TriggerResource or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2020_12_01.models.TriggerResource] + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.models.TriggerResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] @@ -258,7 +258,7 @@ async def get_trigger( :type if_none_match: str :keyword callable cls: A custom type or function that will be passed the direct response :return: TriggerResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.TriggerResource or None + :rtype: ~azure.synapse.artifacts.models.TriggerResource or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.TriggerResource"]] @@ -293,7 +293,7 @@ async def get_trigger( if response.status_code not in [200, 304]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -341,7 +341,7 @@ async def _delete_trigger_initial( if response.status_code not in [200, 202, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -443,7 +443,7 @@ async def _subscribe_trigger_to_events_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -472,7 +472,7 @@ async def begin_subscribe_trigger_to_events( :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of AsyncLROPoller that returns either TriggerSubscriptionOperationStatus or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2020_12_01.models.TriggerSubscriptionOperationStatus] + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.models.TriggerSubscriptionOperationStatus] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] @@ -529,7 +529,7 @@ async def get_event_subscription_status( :type trigger_name: str :keyword callable cls: A custom type or function that will be passed the direct response :return: TriggerSubscriptionOperationStatus, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.TriggerSubscriptionOperationStatus + :rtype: ~azure.synapse.artifacts.models.TriggerSubscriptionOperationStatus :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.TriggerSubscriptionOperationStatus"] @@ -562,7 +562,7 @@ async def get_event_subscription_status( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = self._deserialize('TriggerSubscriptionOperationStatus', pipeline_response) @@ -608,7 +608,7 @@ async def _unsubscribe_trigger_from_events_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -637,7 +637,7 @@ async def begin_unsubscribe_trigger_from_events( :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of AsyncLROPoller that returns either TriggerSubscriptionOperationStatus or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2020_12_01.models.TriggerSubscriptionOperationStatus] + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.models.TriggerSubscriptionOperationStatus] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] @@ -718,7 +718,7 @@ async def _start_trigger_initial( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -820,7 +820,7 @@ async def _stop_trigger_initial( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_trigger_run_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_trigger_run_operations.py similarity index 96% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_trigger_run_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_trigger_run_operations.py index ad5db9f57765..7ca27fd5db14 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_trigger_run_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_trigger_run_operations.py @@ -24,7 +24,7 @@ class TriggerRunOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -87,7 +87,7 @@ async def rerun_trigger_instance( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -143,7 +143,7 @@ async def cancel_trigger_instance( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -159,10 +159,10 @@ async def query_trigger_runs_by_workspace( """Query trigger runs. :param filter_parameters: Parameters to filter the pipeline run. - :type filter_parameters: ~azure.synapse.artifacts.v2020_12_01.models.RunFilterParameters + :type filter_parameters: ~azure.synapse.artifacts.models.RunFilterParameters :keyword callable cls: A custom type or function that will be passed the direct response :return: TriggerRunsQueryResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.TriggerRunsQueryResponse + :rtype: ~azure.synapse.artifacts.models.TriggerRunsQueryResponse :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.TriggerRunsQueryResponse"] @@ -199,7 +199,7 @@ async def query_trigger_runs_by_workspace( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = self._deserialize('TriggerRunsQueryResponse', pipeline_response) diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_workspace_git_repo_management_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_workspace_git_repo_management_operations.py similarity index 96% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_workspace_git_repo_management_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_workspace_git_repo_management_operations.py index ef4ae1cf7a55..09841b358f1a 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_workspace_git_repo_management_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_workspace_git_repo_management_operations.py @@ -24,7 +24,7 @@ class WorkspaceGitRepoManagementOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -48,13 +48,13 @@ async def get_git_hub_access_token( """Get the GitHub access token. :param git_hub_access_token_request: - :type git_hub_access_token_request: ~azure.synapse.artifacts.v2020_12_01.models.GitHubAccessTokenRequest + :type git_hub_access_token_request: ~azure.synapse.artifacts.models.GitHubAccessTokenRequest :param client_request_id: Can provide a guid, which is helpful for debugging and to provide better customer support. :type client_request_id: str :keyword callable cls: A custom type or function that will be passed the direct response :return: GitHubAccessTokenResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.GitHubAccessTokenResponse + :rtype: ~azure.synapse.artifacts.models.GitHubAccessTokenResponse :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.GitHubAccessTokenResponse"] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_workspace_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_workspace_operations.py similarity index 96% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_workspace_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_workspace_operations.py index c0632ef603fd..4a68b69b00a2 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_workspace_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/aio/operations/_workspace_operations.py @@ -24,7 +24,7 @@ class WorkspaceOperations: instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -47,7 +47,7 @@ async def get( :keyword callable cls: A custom type or function that will be passed the direct response :return: Workspace, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.Workspace + :rtype: ~azure.synapse.artifacts.models.Workspace :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.Workspace"] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/models.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/models.py deleted file mode 100644 index e5976174c44e..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/models.py +++ /dev/null @@ -1,7 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -------------------------------------------------------------------------- -from .v2020_12_01.models import * diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/models/__init__.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/models/__init__.py similarity index 96% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/models/__init__.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/models/__init__.py index 72b15de34030..4a88fbd122f0 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/models/__init__.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/models/__init__.py @@ -17,6 +17,13 @@ from ._models_py3 import AmazonMWSLinkedService from ._models_py3 import AmazonMWSObjectDataset from ._models_py3 import AmazonMWSSource + from ._models_py3 import AmazonRdsForOracleLinkedService + from ._models_py3 import AmazonRdsForOraclePartitionSettings + from ._models_py3 import AmazonRdsForOracleSource + from ._models_py3 import AmazonRdsForOracleTableDataset + from ._models_py3 import AmazonRdsForSqlServerLinkedService + from ._models_py3 import AmazonRdsForSqlServerSource + from ._models_py3 import AmazonRdsForSqlServerTableDataset from ._models_py3 import AmazonRedshiftLinkedService from ._models_py3 import AmazonRedshiftSource from ._models_py3 import AmazonRedshiftTableDataset @@ -125,6 +132,7 @@ from ._models_py3 import CassandraTableDataset from ._models_py3 import ChainingTrigger from ._models_py3 import CloudError + from ._models_py3 import CloudErrorAutoGenerated from ._models_py3 import CommonDataServiceForAppsEntityDataset from ._models_py3 import CommonDataServiceForAppsLinkedService from ._models_py3 import CommonDataServiceForAppsSink @@ -163,6 +171,7 @@ from ._models_py3 import DWCopyCommandDefaultValue from ._models_py3 import DWCopyCommandSettings from ._models_py3 import DataFlow + from ._models_py3 import DataFlowDebugCommandPayload from ._models_py3 import DataFlowDebugCommandRequest from ._models_py3 import DataFlowDebugCommandResponse from ._models_py3 import DataFlowDebugPackage @@ -323,6 +332,7 @@ from ._models_py3 import IntegrationRuntimeCustomSetupScriptProperties from ._models_py3 import IntegrationRuntimeDataFlowProperties from ._models_py3 import IntegrationRuntimeDataProxyProperties + from ._models_py3 import IntegrationRuntimeDebugResource from ._models_py3 import IntegrationRuntimeListResponse from ._models_py3 import IntegrationRuntimeReference from ._models_py3 import IntegrationRuntimeResource @@ -338,6 +348,12 @@ from ._models_py3 import JsonSink from ._models_py3 import JsonSource from ._models_py3 import JsonWriteSettings + from ._models_py3 import KqlScript + from ._models_py3 import KqlScriptContent + from ._models_py3 import KqlScriptContentCurrentConnection + from ._models_py3 import KqlScriptContentMetadata + from ._models_py3 import KqlScriptResource + from ._models_py3 import KqlScriptsResourceCollectionResponse from ._models_py3 import LibraryInfo from ._models_py3 import LibraryListResponse from ._models_py3 import LibraryRequirements @@ -395,6 +411,7 @@ from ._models_py3 import Notebook from ._models_py3 import NotebookCell from ._models_py3 import NotebookCellOutputItem + from ._models_py3 import NotebookFolder from ._models_py3 import NotebookKernelSpec from ._models_py3 import NotebookLanguageInfo from ._models_py3 import NotebookListResponse @@ -556,6 +573,7 @@ from ._models_py3 import SparkConfigurationListResponse from ._models_py3 import SparkConfigurationResource from ._models_py3 import SparkJobDefinition + from ._models_py3 import SparkJobDefinitionFolder from ._models_py3 import SparkJobDefinitionResource from ._models_py3 import SparkJobDefinitionsListResponse from ._models_py3 import SparkJobProperties @@ -578,6 +596,7 @@ from ._models_py3 import SqlPoolStoredProcedureActivity from ._models_py3 import SqlScript from ._models_py3 import SqlScriptContent + from ._models_py3 import SqlScriptFolder from ._models_py3 import SqlScriptMetadata from ._models_py3 import SqlScriptResource from ._models_py3 import SqlScriptsListResponse @@ -678,6 +697,13 @@ from ._models import AmazonMWSLinkedService # type: ignore from ._models import AmazonMWSObjectDataset # type: ignore from ._models import AmazonMWSSource # type: ignore + from ._models import AmazonRdsForOracleLinkedService # type: ignore + from ._models import AmazonRdsForOraclePartitionSettings # type: ignore + from ._models import AmazonRdsForOracleSource # type: ignore + from ._models import AmazonRdsForOracleTableDataset # type: ignore + from ._models import AmazonRdsForSqlServerLinkedService # type: ignore + from ._models import AmazonRdsForSqlServerSource # type: ignore + from ._models import AmazonRdsForSqlServerTableDataset # type: ignore from ._models import AmazonRedshiftLinkedService # type: ignore from ._models import AmazonRedshiftSource # type: ignore from ._models import AmazonRedshiftTableDataset # type: ignore @@ -786,6 +812,7 @@ from ._models import CassandraTableDataset # type: ignore from ._models import ChainingTrigger # type: ignore from ._models import CloudError # type: ignore + from ._models import CloudErrorAutoGenerated # type: ignore from ._models import CommonDataServiceForAppsEntityDataset # type: ignore from ._models import CommonDataServiceForAppsLinkedService # type: ignore from ._models import CommonDataServiceForAppsSink # type: ignore @@ -824,6 +851,7 @@ from ._models import DWCopyCommandDefaultValue # type: ignore from ._models import DWCopyCommandSettings # type: ignore from ._models import DataFlow # type: ignore + from ._models import DataFlowDebugCommandPayload # type: ignore from ._models import DataFlowDebugCommandRequest # type: ignore from ._models import DataFlowDebugCommandResponse # type: ignore from ._models import DataFlowDebugPackage # type: ignore @@ -984,6 +1012,7 @@ from ._models import IntegrationRuntimeCustomSetupScriptProperties # type: ignore from ._models import IntegrationRuntimeDataFlowProperties # type: ignore from ._models import IntegrationRuntimeDataProxyProperties # type: ignore + from ._models import IntegrationRuntimeDebugResource # type: ignore from ._models import IntegrationRuntimeListResponse # type: ignore from ._models import IntegrationRuntimeReference # type: ignore from ._models import IntegrationRuntimeResource # type: ignore @@ -999,6 +1028,12 @@ from ._models import JsonSink # type: ignore from ._models import JsonSource # type: ignore from ._models import JsonWriteSettings # type: ignore + from ._models import KqlScript # type: ignore + from ._models import KqlScriptContent # type: ignore + from ._models import KqlScriptContentCurrentConnection # type: ignore + from ._models import KqlScriptContentMetadata # type: ignore + from ._models import KqlScriptResource # type: ignore + from ._models import KqlScriptsResourceCollectionResponse # type: ignore from ._models import LibraryInfo # type: ignore from ._models import LibraryListResponse # type: ignore from ._models import LibraryRequirements # type: ignore @@ -1056,6 +1091,7 @@ from ._models import Notebook # type: ignore from ._models import NotebookCell # type: ignore from ._models import NotebookCellOutputItem # type: ignore + from ._models import NotebookFolder # type: ignore from ._models import NotebookKernelSpec # type: ignore from ._models import NotebookLanguageInfo # type: ignore from ._models import NotebookListResponse # type: ignore @@ -1217,6 +1253,7 @@ from ._models import SparkConfigurationListResponse # type: ignore from ._models import SparkConfigurationResource # type: ignore from ._models import SparkJobDefinition # type: ignore + from ._models import SparkJobDefinitionFolder # type: ignore from ._models import SparkJobDefinitionResource # type: ignore from ._models import SparkJobDefinitionsListResponse # type: ignore from ._models import SparkJobProperties # type: ignore @@ -1239,6 +1276,7 @@ from ._models import SqlPoolStoredProcedureActivity # type: ignore from ._models import SqlScript # type: ignore from ._models import SqlScriptContent # type: ignore + from ._models import SqlScriptFolder # type: ignore from ._models import SqlScriptMetadata # type: ignore from ._models import SqlScriptResource # type: ignore from ._models import SqlScriptsListResponse # type: ignore @@ -1330,6 +1368,7 @@ from ._models import ZohoSource # type: ignore from ._artifacts_client_enums import ( + AmazonRdsForOraclePartitionOption, AvroCompressionCodec, AzureFunctionActivityMethod, AzureSearchIndexWriteBehaviorType, @@ -1340,6 +1379,7 @@ CompressionCodec, CopyBehaviorType, DataFlowComputeType, + DataFlowDebugCommandType, DataFlowReferenceType, DatasetCompressionLevel, DatasetReferenceType, @@ -1348,7 +1388,6 @@ DependencyCondition, DynamicsAuthenticationType, DynamicsDeploymentType, - DynamicsServicePrincipalCredentialType, DynamicsSinkWriteBehavior, EventSubscriptionStatus, ExpressionType, @@ -1372,6 +1411,7 @@ IntegrationRuntimeType, JsonFormatFilePattern, JsonWriteFilePattern, + LivyStates, MongoDbAuthenticationType, NetezzaPartitionOption, NodeSize, @@ -1443,6 +1483,13 @@ 'AmazonMWSLinkedService', 'AmazonMWSObjectDataset', 'AmazonMWSSource', + 'AmazonRdsForOracleLinkedService', + 'AmazonRdsForOraclePartitionSettings', + 'AmazonRdsForOracleSource', + 'AmazonRdsForOracleTableDataset', + 'AmazonRdsForSqlServerLinkedService', + 'AmazonRdsForSqlServerSource', + 'AmazonRdsForSqlServerTableDataset', 'AmazonRedshiftLinkedService', 'AmazonRedshiftSource', 'AmazonRedshiftTableDataset', @@ -1551,6 +1598,7 @@ 'CassandraTableDataset', 'ChainingTrigger', 'CloudError', + 'CloudErrorAutoGenerated', 'CommonDataServiceForAppsEntityDataset', 'CommonDataServiceForAppsLinkedService', 'CommonDataServiceForAppsSink', @@ -1589,6 +1637,7 @@ 'DWCopyCommandDefaultValue', 'DWCopyCommandSettings', 'DataFlow', + 'DataFlowDebugCommandPayload', 'DataFlowDebugCommandRequest', 'DataFlowDebugCommandResponse', 'DataFlowDebugPackage', @@ -1749,6 +1798,7 @@ 'IntegrationRuntimeCustomSetupScriptProperties', 'IntegrationRuntimeDataFlowProperties', 'IntegrationRuntimeDataProxyProperties', + 'IntegrationRuntimeDebugResource', 'IntegrationRuntimeListResponse', 'IntegrationRuntimeReference', 'IntegrationRuntimeResource', @@ -1764,6 +1814,12 @@ 'JsonSink', 'JsonSource', 'JsonWriteSettings', + 'KqlScript', + 'KqlScriptContent', + 'KqlScriptContentCurrentConnection', + 'KqlScriptContentMetadata', + 'KqlScriptResource', + 'KqlScriptsResourceCollectionResponse', 'LibraryInfo', 'LibraryListResponse', 'LibraryRequirements', @@ -1821,6 +1877,7 @@ 'Notebook', 'NotebookCell', 'NotebookCellOutputItem', + 'NotebookFolder', 'NotebookKernelSpec', 'NotebookLanguageInfo', 'NotebookListResponse', @@ -1982,6 +2039,7 @@ 'SparkConfigurationListResponse', 'SparkConfigurationResource', 'SparkJobDefinition', + 'SparkJobDefinitionFolder', 'SparkJobDefinitionResource', 'SparkJobDefinitionsListResponse', 'SparkJobProperties', @@ -2004,6 +2062,7 @@ 'SqlPoolStoredProcedureActivity', 'SqlScript', 'SqlScriptContent', + 'SqlScriptFolder', 'SqlScriptMetadata', 'SqlScriptResource', 'SqlScriptsListResponse', @@ -2093,6 +2152,7 @@ 'ZohoLinkedService', 'ZohoObjectDataset', 'ZohoSource', + 'AmazonRdsForOraclePartitionOption', 'AvroCompressionCodec', 'AzureFunctionActivityMethod', 'AzureSearchIndexWriteBehaviorType', @@ -2103,6 +2163,7 @@ 'CompressionCodec', 'CopyBehaviorType', 'DataFlowComputeType', + 'DataFlowDebugCommandType', 'DataFlowReferenceType', 'DatasetCompressionLevel', 'DatasetReferenceType', @@ -2111,7 +2172,6 @@ 'DependencyCondition', 'DynamicsAuthenticationType', 'DynamicsDeploymentType', - 'DynamicsServicePrincipalCredentialType', 'DynamicsSinkWriteBehavior', 'EventSubscriptionStatus', 'ExpressionType', @@ -2135,6 +2195,7 @@ 'IntegrationRuntimeType', 'JsonFormatFilePattern', 'JsonWriteFilePattern', + 'LivyStates', 'MongoDbAuthenticationType', 'NetezzaPartitionOption', 'NodeSize', diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/models/_artifacts_client_enums.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/models/_artifacts_client_enums.py similarity index 95% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/models/_artifacts_client_enums.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/models/_artifacts_client_enums.py index 5259f055fbc7..467343041d86 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/models/_artifacts_client_enums.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/models/_artifacts_client_enums.py @@ -26,6 +26,12 @@ def __getattr__(cls, name): raise AttributeError(name) +class AmazonRdsForOraclePartitionOption(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): + + NONE = "None" + PHYSICAL_PARTITIONS_OF_TABLE = "PhysicalPartitionsOfTable" + DYNAMIC_RANGE = "DynamicRange" + class AvroCompressionCodec(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): NONE = "none" @@ -92,7 +98,11 @@ class CellOutputType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): ERROR = "error" class CompressionCodec(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): + """All available compressionCodec values. + """ + NONE = "none" + LZO = "lzo" BZIP2 = "bzip2" GZIP = "gzip" DEFLATE = "deflate" @@ -118,6 +128,14 @@ class DataFlowComputeType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): MEMORY_OPTIMIZED = "MemoryOptimized" COMPUTE_OPTIMIZED = "ComputeOptimized" +class DataFlowDebugCommandType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): + """The command type. + """ + + EXECUTE_PREVIEW_QUERY = "executePreviewQuery" + EXECUTE_STATISTICS_QUERY = "executeStatisticsQuery" + EXECUTE_EXPRESSION_QUERY = "executeExpressionQuery" + class DataFlowReferenceType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Data flow reference type. """ @@ -160,9 +178,7 @@ class DependencyCondition(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): COMPLETED = "Completed" class DynamicsAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The authentication type to connect to Dynamics server. 'Office365' for online scenario, 'Ifd' - for on-premises with Ifd scenario, 'AADServicePrincipal' for Server-To-Server authentication in - online scenario. Type: string (or Expression with resultType string). + """All available dynamicsAuthenticationType values. """ OFFICE365 = "Office365" @@ -170,23 +186,12 @@ class DynamicsAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, E AAD_SERVICE_PRINCIPAL = "AADServicePrincipal" class DynamicsDeploymentType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The deployment type of the Dynamics instance. 'Online' for Dynamics Online and - 'OnPremisesWithIfd' for Dynamics on-premises with Ifd. Type: string (or Expression with - resultType string). + """All available dynamicsDeploymentType values. """ ONLINE = "Online" ON_PREMISES_WITH_IFD = "OnPremisesWithIfd" -class DynamicsServicePrincipalCredentialType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The service principal credential type to use in Server-To-Server authentication. - 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' for certificate. Type: string (or - Expression with resultType string). - """ - - SERVICE_PRINCIPAL_KEY = "ServicePrincipalKey" - SERVICE_PRINCIPAL_CERT = "ServicePrincipalCert" - class DynamicsSinkWriteBehavior(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """Defines values for DynamicsSinkWriteBehavior. """ @@ -240,7 +245,7 @@ class HBaseAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum BASIC = "Basic" class HdiNodeTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The node types on which the script action should be executed. + """All available HdiNodeTypes values. """ HEADNODE = "Headnode" @@ -365,13 +370,28 @@ class JsonFormatFilePattern(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)) ARRAY_OF_OBJECTS = "arrayOfObjects" class JsonWriteFilePattern(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """File pattern of JSON. This setting controls the way a collection of JSON objects will be - treated. The default value is 'setOfObjects'. It is case-sensitive. + """All available filePatterns. """ SET_OF_OBJECTS = "setOfObjects" ARRAY_OF_OBJECTS = "arrayOfObjects" +class LivyStates(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): + """The batch state + """ + + NOT_STARTED = "not_started" + STARTING = "starting" + IDLE = "idle" + BUSY = "busy" + SHUTTING_DOWN = "shutting_down" + ERROR = "error" + DEAD = "dead" + KILLED = "killed" + SUCCESS = "success" + RUNNING = "running" + RECOVERING = "recovering" + class MongoDbAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The authentication type to be used to connect to the MongoDB database. """ diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/models/_models.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/models/_models.py similarity index 91% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/models/_models.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/models/_models.py index 2443cbf2994c..472566bda3da 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/models/_models.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/models/_models.py @@ -28,9 +28,9 @@ class Activity(msrest.serialization.Model): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] """ _validation = { @@ -75,8 +75,7 @@ class ActivityDependency(msrest.serialization.Model): :param activity: Required. Activity name. :type activity: str :param dependency_conditions: Required. Match-Condition for the dependency. - :type dependency_conditions: list[str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DependencyCondition] + :type dependency_conditions: list[str or ~azure.synapse.artifacts.models.DependencyCondition] """ _validation = { @@ -246,7 +245,7 @@ class ActivityRunsQueryResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of activity runs. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityRun] + :type value: list[~azure.synapse.artifacts.models.ActivityRun] :param continuation_token: The continuation token for getting the next page of results, if any remaining results exist, null otherwise. :type continuation_token: str @@ -316,7 +315,7 @@ class LinkedService(msrest.serialization.Model): """The Azure Synapse nested object which contains the information and credential which can be used to connect with related store or compute resource. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonMWSLinkedService, AmazonRedshiftLinkedService, AmazonS3LinkedService, AzureBatchLinkedService, AzureBlobFSLinkedService, AzureBlobStorageLinkedService, AzureDataExplorerLinkedService, AzureDataLakeAnalyticsLinkedService, AzureDataLakeStoreLinkedService, AzureDatabricksLinkedService, AzureDatabricksDeltaLakeLinkedService, AzureFileStorageLinkedService, AzureFunctionLinkedService, AzureKeyVaultLinkedService, AzureMLLinkedService, AzureMLServiceLinkedService, AzureMariaDBLinkedService, AzureMySqlLinkedService, AzurePostgreSqlLinkedService, AzureSearchLinkedService, AzureSqlDWLinkedService, AzureSqlDatabaseLinkedService, AzureSqlMILinkedService, AzureStorageLinkedService, AzureTableStorageLinkedService, CassandraLinkedService, CommonDataServiceForAppsLinkedService, ConcurLinkedService, CosmosDbLinkedService, CosmosDbMongoDbApiLinkedService, CouchbaseLinkedService, CustomDataSourceLinkedService, Db2LinkedService, DrillLinkedService, DynamicsLinkedService, DynamicsAXLinkedService, DynamicsCrmLinkedService, EloquaLinkedService, FileServerLinkedService, FtpServerLinkedService, GoogleAdWordsLinkedService, GoogleBigQueryLinkedService, GoogleCloudStorageLinkedService, GreenplumLinkedService, HBaseLinkedService, HDInsightLinkedService, HDInsightOnDemandLinkedService, HdfsLinkedService, HiveLinkedService, HttpLinkedService, HubspotLinkedService, ImpalaLinkedService, InformixLinkedService, JiraLinkedService, MagentoLinkedService, MariaDBLinkedService, MarketoLinkedService, MicrosoftAccessLinkedService, MongoDbLinkedService, MongoDbAtlasLinkedService, MongoDbV2LinkedService, MySqlLinkedService, NetezzaLinkedService, ODataLinkedService, OdbcLinkedService, Office365LinkedService, OracleLinkedService, OracleServiceCloudLinkedService, PaypalLinkedService, PhoenixLinkedService, PostgreSqlLinkedService, PrestoLinkedService, QuickBooksLinkedService, ResponsysLinkedService, RestServiceLinkedService, SalesforceLinkedService, SalesforceMarketingCloudLinkedService, SalesforceServiceCloudLinkedService, SapBWLinkedService, SapCloudForCustomerLinkedService, SapEccLinkedService, SapHanaLinkedService, SapOpenHubLinkedService, SapTableLinkedService, ServiceNowLinkedService, SftpServerLinkedService, SharePointOnlineListLinkedService, ShopifyLinkedService, SnowflakeLinkedService, SparkLinkedService, SqlServerLinkedService, SquareLinkedService, SybaseLinkedService, TeradataLinkedService, VerticaLinkedService, WebLinkedService, XeroLinkedService, ZohoLinkedService. + sub-classes are: AmazonMWSLinkedService, AmazonRdsForOracleLinkedService, AmazonRdsForSqlServerLinkedService, AmazonRedshiftLinkedService, AmazonS3LinkedService, AzureBatchLinkedService, AzureBlobFSLinkedService, AzureBlobStorageLinkedService, AzureDataExplorerLinkedService, AzureDataLakeAnalyticsLinkedService, AzureDataLakeStoreLinkedService, AzureDatabricksLinkedService, AzureDatabricksDeltaLakeLinkedService, AzureFileStorageLinkedService, AzureFunctionLinkedService, AzureKeyVaultLinkedService, AzureMLLinkedService, AzureMLServiceLinkedService, AzureMariaDBLinkedService, AzureMySqlLinkedService, AzurePostgreSqlLinkedService, AzureSearchLinkedService, AzureSqlDWLinkedService, AzureSqlDatabaseLinkedService, AzureSqlMILinkedService, AzureStorageLinkedService, AzureTableStorageLinkedService, CassandraLinkedService, CommonDataServiceForAppsLinkedService, ConcurLinkedService, CosmosDbLinkedService, CosmosDbMongoDbApiLinkedService, CouchbaseLinkedService, CustomDataSourceLinkedService, Db2LinkedService, DrillLinkedService, DynamicsLinkedService, DynamicsAXLinkedService, DynamicsCrmLinkedService, EloquaLinkedService, FileServerLinkedService, FtpServerLinkedService, GoogleAdWordsLinkedService, GoogleBigQueryLinkedService, GoogleCloudStorageLinkedService, GreenplumLinkedService, HBaseLinkedService, HDInsightLinkedService, HDInsightOnDemandLinkedService, HdfsLinkedService, HiveLinkedService, HttpLinkedService, HubspotLinkedService, ImpalaLinkedService, InformixLinkedService, JiraLinkedService, MagentoLinkedService, MariaDBLinkedService, MarketoLinkedService, MicrosoftAccessLinkedService, MongoDbLinkedService, MongoDbAtlasLinkedService, MongoDbV2LinkedService, MySqlLinkedService, NetezzaLinkedService, ODataLinkedService, OdbcLinkedService, Office365LinkedService, OracleLinkedService, OracleServiceCloudLinkedService, PaypalLinkedService, PhoenixLinkedService, PostgreSqlLinkedService, PrestoLinkedService, QuickBooksLinkedService, ResponsysLinkedService, RestServiceLinkedService, SalesforceLinkedService, SalesforceMarketingCloudLinkedService, SalesforceServiceCloudLinkedService, SapBWLinkedService, SapCloudForCustomerLinkedService, SapEccLinkedService, SapHanaLinkedService, SapOpenHubLinkedService, SapTableLinkedService, ServiceNowLinkedService, SftpServerLinkedService, SharePointOnlineListLinkedService, ShopifyLinkedService, SnowflakeLinkedService, SparkLinkedService, SqlServerLinkedService, SquareLinkedService, SybaseLinkedService, TeradataLinkedService, VerticaLinkedService, WebLinkedService, XeroLinkedService, ZohoLinkedService. All required parameters must be populated in order to send to Azure. @@ -326,13 +325,11 @@ class LinkedService(msrest.serialization.Model): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] """ @@ -351,7 +348,7 @@ class LinkedService(msrest.serialization.Model): } _subtype_map = { - 'type': {'AmazonMWS': 'AmazonMWSLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureBlobFS': 'AzureBlobFSLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureDataExplorer': 'AzureDataExplorerLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDatabricksDeltaLake': 'AzureDatabricksDeltaLakeLinkedService', 'AzureFileStorage': 'AzureFileStorageLinkedService', 'AzureFunction': 'AzureFunctionLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureML': 'AzureMLLinkedService', 'AzureMLService': 'AzureMLServiceLinkedService', 'AzureMariaDB': 'AzureMariaDBLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'AzureSqlMI': 'AzureSqlMILinkedService', 'AzureStorage': 'AzureStorageLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'Cassandra': 'CassandraLinkedService', 'CommonDataServiceForApps': 'CommonDataServiceForAppsLinkedService', 'Concur': 'ConcurLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'CosmosDbMongoDbApi': 'CosmosDbMongoDbApiLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'Db2': 'Db2LinkedService', 'Drill': 'DrillLinkedService', 'Dynamics': 'DynamicsLinkedService', 'DynamicsAX': 'DynamicsAXLinkedService', 'DynamicsCrm': 'DynamicsCrmLinkedService', 'Eloqua': 'EloquaLinkedService', 'FileServer': 'FileServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'GoogleAdWords': 'GoogleAdWordsLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'GoogleCloudStorage': 'GoogleCloudStorageLinkedService', 'Greenplum': 'GreenplumLinkedService', 'HBase': 'HBaseLinkedService', 'HDInsight': 'HDInsightLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'Hdfs': 'HdfsLinkedService', 'Hive': 'HiveLinkedService', 'HttpServer': 'HttpLinkedService', 'Hubspot': 'HubspotLinkedService', 'Impala': 'ImpalaLinkedService', 'Informix': 'InformixLinkedService', 'Jira': 'JiraLinkedService', 'Magento': 'MagentoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Marketo': 'MarketoLinkedService', 'MicrosoftAccess': 'MicrosoftAccessLinkedService', 'MongoDb': 'MongoDbLinkedService', 'MongoDbAtlas': 'MongoDbAtlasLinkedService', 'MongoDbV2': 'MongoDbV2LinkedService', 'MySql': 'MySqlLinkedService', 'Netezza': 'NetezzaLinkedService', 'OData': 'ODataLinkedService', 'Odbc': 'OdbcLinkedService', 'Office365': 'Office365LinkedService', 'Oracle': 'OracleLinkedService', 'OracleServiceCloud': 'OracleServiceCloudLinkedService', 'Paypal': 'PaypalLinkedService', 'Phoenix': 'PhoenixLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'Presto': 'PrestoLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Responsys': 'ResponsysLinkedService', 'RestService': 'RestServiceLinkedService', 'Salesforce': 'SalesforceLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'SalesforceServiceCloud': 'SalesforceServiceCloudLinkedService', 'SapBW': 'SapBWLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'SapEcc': 'SapEccLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapOpenHub': 'SapOpenHubLinkedService', 'SapTable': 'SapTableLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'Sftp': 'SftpServerLinkedService', 'SharePointOnlineList': 'SharePointOnlineListLinkedService', 'Shopify': 'ShopifyLinkedService', 'Snowflake': 'SnowflakeLinkedService', 'Spark': 'SparkLinkedService', 'SqlServer': 'SqlServerLinkedService', 'Square': 'SquareLinkedService', 'Sybase': 'SybaseLinkedService', 'Teradata': 'TeradataLinkedService', 'Vertica': 'VerticaLinkedService', 'Web': 'WebLinkedService', 'Xero': 'XeroLinkedService', 'Zoho': 'ZohoLinkedService'} + 'type': {'AmazonMWS': 'AmazonMWSLinkedService', 'AmazonRdsForOracle': 'AmazonRdsForOracleLinkedService', 'AmazonRdsForSqlServer': 'AmazonRdsForSqlServerLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureBlobFS': 'AzureBlobFSLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureDataExplorer': 'AzureDataExplorerLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDatabricksDeltaLake': 'AzureDatabricksDeltaLakeLinkedService', 'AzureFileStorage': 'AzureFileStorageLinkedService', 'AzureFunction': 'AzureFunctionLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureML': 'AzureMLLinkedService', 'AzureMLService': 'AzureMLServiceLinkedService', 'AzureMariaDB': 'AzureMariaDBLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'AzureSqlMI': 'AzureSqlMILinkedService', 'AzureStorage': 'AzureStorageLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'Cassandra': 'CassandraLinkedService', 'CommonDataServiceForApps': 'CommonDataServiceForAppsLinkedService', 'Concur': 'ConcurLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'CosmosDbMongoDbApi': 'CosmosDbMongoDbApiLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'Db2': 'Db2LinkedService', 'Drill': 'DrillLinkedService', 'Dynamics': 'DynamicsLinkedService', 'DynamicsAX': 'DynamicsAXLinkedService', 'DynamicsCrm': 'DynamicsCrmLinkedService', 'Eloqua': 'EloquaLinkedService', 'FileServer': 'FileServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'GoogleAdWords': 'GoogleAdWordsLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'GoogleCloudStorage': 'GoogleCloudStorageLinkedService', 'Greenplum': 'GreenplumLinkedService', 'HBase': 'HBaseLinkedService', 'HDInsight': 'HDInsightLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'Hdfs': 'HdfsLinkedService', 'Hive': 'HiveLinkedService', 'HttpServer': 'HttpLinkedService', 'Hubspot': 'HubspotLinkedService', 'Impala': 'ImpalaLinkedService', 'Informix': 'InformixLinkedService', 'Jira': 'JiraLinkedService', 'Magento': 'MagentoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Marketo': 'MarketoLinkedService', 'MicrosoftAccess': 'MicrosoftAccessLinkedService', 'MongoDb': 'MongoDbLinkedService', 'MongoDbAtlas': 'MongoDbAtlasLinkedService', 'MongoDbV2': 'MongoDbV2LinkedService', 'MySql': 'MySqlLinkedService', 'Netezza': 'NetezzaLinkedService', 'OData': 'ODataLinkedService', 'Odbc': 'OdbcLinkedService', 'Office365': 'Office365LinkedService', 'Oracle': 'OracleLinkedService', 'OracleServiceCloud': 'OracleServiceCloudLinkedService', 'Paypal': 'PaypalLinkedService', 'Phoenix': 'PhoenixLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'Presto': 'PrestoLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Responsys': 'ResponsysLinkedService', 'RestService': 'RestServiceLinkedService', 'Salesforce': 'SalesforceLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'SalesforceServiceCloud': 'SalesforceServiceCloudLinkedService', 'SapBW': 'SapBWLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'SapEcc': 'SapEccLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapOpenHub': 'SapOpenHubLinkedService', 'SapTable': 'SapTableLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'Sftp': 'SftpServerLinkedService', 'SharePointOnlineList': 'SharePointOnlineListLinkedService', 'Shopify': 'ShopifyLinkedService', 'Snowflake': 'SnowflakeLinkedService', 'Spark': 'SparkLinkedService', 'SqlServer': 'SqlServerLinkedService', 'Square': 'SquareLinkedService', 'Sybase': 'SybaseLinkedService', 'Teradata': 'TeradataLinkedService', 'Vertica': 'VerticaLinkedService', 'Web': 'WebLinkedService', 'Xero': 'XeroLinkedService', 'Zoho': 'ZohoLinkedService'} } def __init__( @@ -378,13 +375,11 @@ class AmazonMWSLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param endpoint: Required. The endpoint of the Amazon MWS server, (i.e. @@ -397,11 +392,11 @@ class AmazonMWSLinkedService(LinkedService): :param seller_id: Required. The Amazon seller ID. :type seller_id: any :param mws_auth_token: The Amazon MWS authentication token. - :type mws_auth_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type mws_auth_token: ~azure.synapse.artifacts.models.SecretBase :param access_key_id: Required. The access key id used to access data. :type access_key_id: any :param secret_key: The secret key used to access data. - :type secret_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type secret_key: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -467,7 +462,7 @@ class Dataset(msrest.serialization.Model): """The Azure Data Factory nested object which identifies data within different data stores, such as tables, files, folders, and documents. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonMWSObjectDataset, AmazonRedshiftTableDataset, AmazonS3Dataset, AvroDataset, AzureBlobDataset, AzureBlobFSDataset, AzureDataExplorerTableDataset, AzureDataLakeStoreDataset, AzureDatabricksDeltaLakeDataset, AzureMariaDBTableDataset, AzureMySqlTableDataset, AzurePostgreSqlTableDataset, AzureSearchIndexDataset, AzureSqlDWTableDataset, AzureSqlMITableDataset, AzureSqlTableDataset, AzureTableDataset, BinaryDataset, CassandraTableDataset, CommonDataServiceForAppsEntityDataset, ConcurObjectDataset, CosmosDbMongoDbApiCollectionDataset, CosmosDbSqlApiCollectionDataset, CouchbaseTableDataset, CustomDataset, Db2TableDataset, DelimitedTextDataset, DocumentDbCollectionDataset, DrillTableDataset, DynamicsAXResourceDataset, DynamicsCrmEntityDataset, DynamicsEntityDataset, EloquaObjectDataset, ExcelDataset, FileShareDataset, GoogleAdWordsObjectDataset, GoogleBigQueryObjectDataset, GreenplumTableDataset, HBaseObjectDataset, HiveObjectDataset, HttpDataset, HubspotObjectDataset, ImpalaObjectDataset, InformixTableDataset, JiraObjectDataset, JsonDataset, MagentoObjectDataset, MariaDBTableDataset, MarketoObjectDataset, MicrosoftAccessTableDataset, MongoDbAtlasCollectionDataset, MongoDbCollectionDataset, MongoDbV2CollectionDataset, MySqlTableDataset, NetezzaTableDataset, ODataResourceDataset, OdbcTableDataset, Office365Dataset, OracleServiceCloudObjectDataset, OracleTableDataset, OrcDataset, ParquetDataset, PaypalObjectDataset, PhoenixObjectDataset, PostgreSqlTableDataset, PrestoObjectDataset, QuickBooksObjectDataset, RelationalTableDataset, ResponsysObjectDataset, RestResourceDataset, SalesforceMarketingCloudObjectDataset, SalesforceObjectDataset, SalesforceServiceCloudObjectDataset, SapBwCubeDataset, SapCloudForCustomerResourceDataset, SapEccResourceDataset, SapHanaTableDataset, SapOpenHubTableDataset, SapTableResourceDataset, ServiceNowObjectDataset, SharePointOnlineListResourceDataset, ShopifyObjectDataset, SnowflakeDataset, SparkObjectDataset, SqlServerTableDataset, SquareObjectDataset, SybaseTableDataset, TeradataTableDataset, VerticaTableDataset, WebTableDataset, XeroObjectDataset, XmlDataset, ZohoObjectDataset. + sub-classes are: AmazonMWSObjectDataset, AmazonRdsForOracleTableDataset, AmazonRdsForSqlServerTableDataset, AmazonRedshiftTableDataset, AmazonS3Dataset, AvroDataset, AzureBlobDataset, AzureBlobFSDataset, AzureDataExplorerTableDataset, AzureDataLakeStoreDataset, AzureDatabricksDeltaLakeDataset, AzureMariaDBTableDataset, AzureMySqlTableDataset, AzurePostgreSqlTableDataset, AzureSearchIndexDataset, AzureSqlDWTableDataset, AzureSqlMITableDataset, AzureSqlTableDataset, AzureTableDataset, BinaryDataset, CassandraTableDataset, CommonDataServiceForAppsEntityDataset, ConcurObjectDataset, CosmosDbMongoDbApiCollectionDataset, CosmosDbSqlApiCollectionDataset, CouchbaseTableDataset, CustomDataset, Db2TableDataset, DelimitedTextDataset, DocumentDbCollectionDataset, DrillTableDataset, DynamicsAXResourceDataset, DynamicsCrmEntityDataset, DynamicsEntityDataset, EloquaObjectDataset, ExcelDataset, FileShareDataset, GoogleAdWordsObjectDataset, GoogleBigQueryObjectDataset, GreenplumTableDataset, HBaseObjectDataset, HiveObjectDataset, HttpDataset, HubspotObjectDataset, ImpalaObjectDataset, InformixTableDataset, JiraObjectDataset, JsonDataset, MagentoObjectDataset, MariaDBTableDataset, MarketoObjectDataset, MicrosoftAccessTableDataset, MongoDbAtlasCollectionDataset, MongoDbCollectionDataset, MongoDbV2CollectionDataset, MySqlTableDataset, NetezzaTableDataset, ODataResourceDataset, OdbcTableDataset, Office365Dataset, OracleServiceCloudObjectDataset, OracleTableDataset, OrcDataset, ParquetDataset, PaypalObjectDataset, PhoenixObjectDataset, PostgreSqlTableDataset, PrestoObjectDataset, QuickBooksObjectDataset, RelationalTableDataset, ResponsysObjectDataset, RestResourceDataset, SalesforceMarketingCloudObjectDataset, SalesforceObjectDataset, SalesforceServiceCloudObjectDataset, SapBwCubeDataset, SapCloudForCustomerResourceDataset, SapEccResourceDataset, SapHanaTableDataset, SapOpenHubTableDataset, SapTableResourceDataset, ServiceNowObjectDataset, SharePointOnlineListResourceDataset, ShopifyObjectDataset, SnowflakeDataset, SparkObjectDataset, SqlServerTableDataset, SquareObjectDataset, SybaseTableDataset, TeradataTableDataset, VerticaTableDataset, WebTableDataset, XeroObjectDataset, XmlDataset, ZohoObjectDataset. All required parameters must be populated in order to send to Azure. @@ -485,16 +480,14 @@ class Dataset(msrest.serialization.Model): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder """ _validation = { @@ -515,7 +508,7 @@ class Dataset(msrest.serialization.Model): } _subtype_map = { - 'type': {'AmazonMWSObject': 'AmazonMWSObjectDataset', 'AmazonRedshiftTable': 'AmazonRedshiftTableDataset', 'AmazonS3Object': 'AmazonS3Dataset', 'Avro': 'AvroDataset', 'AzureBlob': 'AzureBlobDataset', 'AzureBlobFSFile': 'AzureBlobFSDataset', 'AzureDataExplorerTable': 'AzureDataExplorerTableDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'AzureDatabricksDeltaLakeDataset': 'AzureDatabricksDeltaLakeDataset', 'AzureMariaDBTable': 'AzureMariaDBTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlMITable': 'AzureSqlMITableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'Binary': 'BinaryDataset', 'CassandraTable': 'CassandraTableDataset', 'CommonDataServiceForAppsEntity': 'CommonDataServiceForAppsEntityDataset', 'ConcurObject': 'ConcurObjectDataset', 'CosmosDbMongoDbApiCollection': 'CosmosDbMongoDbApiCollectionDataset', 'CosmosDbSqlApiCollection': 'CosmosDbSqlApiCollectionDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'CustomDataset': 'CustomDataset', 'Db2Table': 'Db2TableDataset', 'DelimitedText': 'DelimitedTextDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'DrillTable': 'DrillTableDataset', 'DynamicsAXResource': 'DynamicsAXResourceDataset', 'DynamicsCrmEntity': 'DynamicsCrmEntityDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'EloquaObject': 'EloquaObjectDataset', 'Excel': 'ExcelDataset', 'FileShare': 'FileShareDataset', 'GoogleAdWordsObject': 'GoogleAdWordsObjectDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'HBaseObject': 'HBaseObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HttpFile': 'HttpDataset', 'HubspotObject': 'HubspotObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'InformixTable': 'InformixTableDataset', 'JiraObject': 'JiraObjectDataset', 'Json': 'JsonDataset', 'MagentoObject': 'MagentoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MarketoObject': 'MarketoObjectDataset', 'MicrosoftAccessTable': 'MicrosoftAccessTableDataset', 'MongoDbAtlasCollection': 'MongoDbAtlasCollectionDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'MongoDbV2Collection': 'MongoDbV2CollectionDataset', 'MySqlTable': 'MySqlTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ODataResource': 'ODataResourceDataset', 'OdbcTable': 'OdbcTableDataset', 'Office365Table': 'Office365Dataset', 'OracleServiceCloudObject': 'OracleServiceCloudObjectDataset', 'OracleTable': 'OracleTableDataset', 'Orc': 'OrcDataset', 'Parquet': 'ParquetDataset', 'PaypalObject': 'PaypalObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PostgreSqlTable': 'PostgreSqlTableDataset', 'PrestoObject': 'PrestoObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'ResponsysObject': 'ResponsysObjectDataset', 'RestResource': 'RestResourceDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'SalesforceServiceCloudObject': 'SalesforceServiceCloudObjectDataset', 'SapBwCube': 'SapBwCubeDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapHanaTable': 'SapHanaTableDataset', 'SapOpenHubTable': 'SapOpenHubTableDataset', 'SapTableResource': 'SapTableResourceDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'SharePointOnlineListResource': 'SharePointOnlineListResourceDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'SnowflakeTable': 'SnowflakeDataset', 'SparkObject': 'SparkObjectDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SquareObject': 'SquareObjectDataset', 'SybaseTable': 'SybaseTableDataset', 'TeradataTable': 'TeradataTableDataset', 'VerticaTable': 'VerticaTableDataset', 'WebTable': 'WebTableDataset', 'XeroObject': 'XeroObjectDataset', 'Xml': 'XmlDataset', 'ZohoObject': 'ZohoObjectDataset'} + 'type': {'AmazonMWSObject': 'AmazonMWSObjectDataset', 'AmazonRdsForOracleTable': 'AmazonRdsForOracleTableDataset', 'AmazonRdsForSqlServerTable': 'AmazonRdsForSqlServerTableDataset', 'AmazonRedshiftTable': 'AmazonRedshiftTableDataset', 'AmazonS3Object': 'AmazonS3Dataset', 'Avro': 'AvroDataset', 'AzureBlob': 'AzureBlobDataset', 'AzureBlobFSFile': 'AzureBlobFSDataset', 'AzureDataExplorerTable': 'AzureDataExplorerTableDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'AzureDatabricksDeltaLakeDataset': 'AzureDatabricksDeltaLakeDataset', 'AzureMariaDBTable': 'AzureMariaDBTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlMITable': 'AzureSqlMITableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'Binary': 'BinaryDataset', 'CassandraTable': 'CassandraTableDataset', 'CommonDataServiceForAppsEntity': 'CommonDataServiceForAppsEntityDataset', 'ConcurObject': 'ConcurObjectDataset', 'CosmosDbMongoDbApiCollection': 'CosmosDbMongoDbApiCollectionDataset', 'CosmosDbSqlApiCollection': 'CosmosDbSqlApiCollectionDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'CustomDataset': 'CustomDataset', 'Db2Table': 'Db2TableDataset', 'DelimitedText': 'DelimitedTextDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'DrillTable': 'DrillTableDataset', 'DynamicsAXResource': 'DynamicsAXResourceDataset', 'DynamicsCrmEntity': 'DynamicsCrmEntityDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'EloquaObject': 'EloquaObjectDataset', 'Excel': 'ExcelDataset', 'FileShare': 'FileShareDataset', 'GoogleAdWordsObject': 'GoogleAdWordsObjectDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'HBaseObject': 'HBaseObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HttpFile': 'HttpDataset', 'HubspotObject': 'HubspotObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'InformixTable': 'InformixTableDataset', 'JiraObject': 'JiraObjectDataset', 'Json': 'JsonDataset', 'MagentoObject': 'MagentoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MarketoObject': 'MarketoObjectDataset', 'MicrosoftAccessTable': 'MicrosoftAccessTableDataset', 'MongoDbAtlasCollection': 'MongoDbAtlasCollectionDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'MongoDbV2Collection': 'MongoDbV2CollectionDataset', 'MySqlTable': 'MySqlTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ODataResource': 'ODataResourceDataset', 'OdbcTable': 'OdbcTableDataset', 'Office365Table': 'Office365Dataset', 'OracleServiceCloudObject': 'OracleServiceCloudObjectDataset', 'OracleTable': 'OracleTableDataset', 'Orc': 'OrcDataset', 'Parquet': 'ParquetDataset', 'PaypalObject': 'PaypalObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PostgreSqlTable': 'PostgreSqlTableDataset', 'PrestoObject': 'PrestoObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'ResponsysObject': 'ResponsysObjectDataset', 'RestResource': 'RestResourceDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'SalesforceServiceCloudObject': 'SalesforceServiceCloudObjectDataset', 'SapBwCube': 'SapBwCubeDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapHanaTable': 'SapHanaTableDataset', 'SapOpenHubTable': 'SapOpenHubTableDataset', 'SapTableResource': 'SapTableResourceDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'SharePointOnlineListResource': 'SharePointOnlineListResourceDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'SnowflakeTable': 'SnowflakeDataset', 'SparkObject': 'SparkObjectDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SquareObject': 'SquareObjectDataset', 'SybaseTable': 'SybaseTableDataset', 'TeradataTable': 'TeradataTableDataset', 'VerticaTable': 'VerticaTableDataset', 'WebTable': 'WebTableDataset', 'XeroObject': 'XeroObjectDataset', 'Xml': 'XmlDataset', 'ZohoObject': 'ZohoObjectDataset'} } def __init__( @@ -553,16 +546,14 @@ class AmazonMWSObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -598,7 +589,7 @@ class CopySource(msrest.serialization.Model): """A copy activity source. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AvroSource, AzureBlobFSSource, AzureDataExplorerSource, AzureDataLakeStoreSource, AzureDatabricksDeltaLakeSource, BinarySource, BlobSource, CommonDataServiceForAppsSource, CosmosDbMongoDbApiSource, CosmosDbSqlApiSource, DelimitedTextSource, DocumentDbCollectionSource, DynamicsCrmSource, DynamicsSource, ExcelSource, FileSystemSource, HdfsSource, HttpSource, JsonSource, MicrosoftAccessSource, MongoDbAtlasSource, MongoDbSource, MongoDbV2Source, ODataSource, Office365Source, OracleSource, OrcSource, ParquetSource, RelationalSource, RestSource, SalesforceServiceCloudSource, SharePointOnlineListSource, SnowflakeSource, TabularSource, WebSource, XmlSource. + sub-classes are: AmazonRdsForOracleSource, AvroSource, AzureBlobFSSource, AzureDataExplorerSource, AzureDataLakeStoreSource, AzureDatabricksDeltaLakeSource, BinarySource, BlobSource, CommonDataServiceForAppsSource, CosmosDbMongoDbApiSource, CosmosDbSqlApiSource, DelimitedTextSource, DocumentDbCollectionSource, DynamicsCrmSource, DynamicsSource, ExcelSource, FileSystemSource, HdfsSource, HttpSource, JsonSource, MicrosoftAccessSource, MongoDbAtlasSource, MongoDbSource, MongoDbV2Source, ODataSource, Office365Source, OracleSource, OrcSource, ParquetSource, RelationalSource, RestSource, SalesforceServiceCloudSource, SharePointOnlineListSource, SnowflakeSource, TabularSource, WebSource, XmlSource. All required parameters must be populated in order to send to Azure. @@ -631,7 +622,7 @@ class CopySource(msrest.serialization.Model): } _subtype_map = { - 'type': {'AvroSource': 'AvroSource', 'AzureBlobFSSource': 'AzureBlobFSSource', 'AzureDataExplorerSource': 'AzureDataExplorerSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'AzureDatabricksDeltaLakeSource': 'AzureDatabricksDeltaLakeSource', 'BinarySource': 'BinarySource', 'BlobSource': 'BlobSource', 'CommonDataServiceForAppsSource': 'CommonDataServiceForAppsSource', 'CosmosDbMongoDbApiSource': 'CosmosDbMongoDbApiSource', 'CosmosDbSqlApiSource': 'CosmosDbSqlApiSource', 'DelimitedTextSource': 'DelimitedTextSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'DynamicsCrmSource': 'DynamicsCrmSource', 'DynamicsSource': 'DynamicsSource', 'ExcelSource': 'ExcelSource', 'FileSystemSource': 'FileSystemSource', 'HdfsSource': 'HdfsSource', 'HttpSource': 'HttpSource', 'JsonSource': 'JsonSource', 'MicrosoftAccessSource': 'MicrosoftAccessSource', 'MongoDbAtlasSource': 'MongoDbAtlasSource', 'MongoDbSource': 'MongoDbSource', 'MongoDbV2Source': 'MongoDbV2Source', 'ODataSource': 'ODataSource', 'Office365Source': 'Office365Source', 'OracleSource': 'OracleSource', 'OrcSource': 'OrcSource', 'ParquetSource': 'ParquetSource', 'RelationalSource': 'RelationalSource', 'RestSource': 'RestSource', 'SalesforceServiceCloudSource': 'SalesforceServiceCloudSource', 'SharePointOnlineListSource': 'SharePointOnlineListSource', 'SnowflakeSource': 'SnowflakeSource', 'TabularSource': 'TabularSource', 'WebSource': 'WebSource', 'XmlSource': 'XmlSource'} + 'type': {'AmazonRdsForOracleSource': 'AmazonRdsForOracleSource', 'AvroSource': 'AvroSource', 'AzureBlobFSSource': 'AzureBlobFSSource', 'AzureDataExplorerSource': 'AzureDataExplorerSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'AzureDatabricksDeltaLakeSource': 'AzureDatabricksDeltaLakeSource', 'BinarySource': 'BinarySource', 'BlobSource': 'BlobSource', 'CommonDataServiceForAppsSource': 'CommonDataServiceForAppsSource', 'CosmosDbMongoDbApiSource': 'CosmosDbMongoDbApiSource', 'CosmosDbSqlApiSource': 'CosmosDbSqlApiSource', 'DelimitedTextSource': 'DelimitedTextSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'DynamicsCrmSource': 'DynamicsCrmSource', 'DynamicsSource': 'DynamicsSource', 'ExcelSource': 'ExcelSource', 'FileSystemSource': 'FileSystemSource', 'HdfsSource': 'HdfsSource', 'HttpSource': 'HttpSource', 'JsonSource': 'JsonSource', 'MicrosoftAccessSource': 'MicrosoftAccessSource', 'MongoDbAtlasSource': 'MongoDbAtlasSource', 'MongoDbSource': 'MongoDbSource', 'MongoDbV2Source': 'MongoDbV2Source', 'ODataSource': 'ODataSource', 'Office365Source': 'Office365Source', 'OracleSource': 'OracleSource', 'OrcSource': 'OrcSource', 'ParquetSource': 'ParquetSource', 'RelationalSource': 'RelationalSource', 'RestSource': 'RestSource', 'SalesforceServiceCloudSource': 'SalesforceServiceCloudSource', 'SharePointOnlineListSource': 'SharePointOnlineListSource', 'SnowflakeSource': 'SnowflakeSource', 'TabularSource': 'TabularSource', 'WebSource': 'WebSource', 'XmlSource': 'XmlSource'} } def __init__( @@ -650,7 +641,7 @@ class TabularSource(CopySource): """Copy activity sources of tabular type. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonMWSSource, AmazonRedshiftSource, AzureMariaDBSource, AzureMySqlSource, AzurePostgreSqlSource, AzureSqlSource, AzureTableSource, CassandraSource, ConcurSource, CouchbaseSource, Db2Source, DrillSource, DynamicsAXSource, EloquaSource, GoogleAdWordsSource, GoogleBigQuerySource, GreenplumSource, HBaseSource, HiveSource, HubspotSource, ImpalaSource, InformixSource, JiraSource, MagentoSource, MariaDBSource, MarketoSource, MySqlSource, NetezzaSource, OdbcSource, OracleServiceCloudSource, PaypalSource, PhoenixSource, PostgreSqlSource, PrestoSource, QuickBooksSource, ResponsysSource, SalesforceMarketingCloudSource, SalesforceSource, SapBwSource, SapCloudForCustomerSource, SapEccSource, SapHanaSource, SapOpenHubSource, SapTableSource, ServiceNowSource, ShopifySource, SparkSource, SqlDWSource, SqlMISource, SqlServerSource, SqlSource, SquareSource, SybaseSource, TeradataSource, VerticaSource, XeroSource, ZohoSource. + sub-classes are: AmazonMWSSource, AmazonRdsForSqlServerSource, AmazonRedshiftSource, AzureMariaDBSource, AzureMySqlSource, AzurePostgreSqlSource, AzureSqlSource, AzureTableSource, CassandraSource, ConcurSource, CouchbaseSource, Db2Source, DrillSource, DynamicsAXSource, EloquaSource, GoogleAdWordsSource, GoogleBigQuerySource, GreenplumSource, HBaseSource, HiveSource, HubspotSource, ImpalaSource, InformixSource, JiraSource, MagentoSource, MariaDBSource, MarketoSource, MySqlSource, NetezzaSource, OdbcSource, OracleServiceCloudSource, PaypalSource, PhoenixSource, PostgreSqlSource, PrestoSource, QuickBooksSource, ResponsysSource, SalesforceMarketingCloudSource, SalesforceSource, SapBwSource, SapCloudForCustomerSource, SapEccSource, SapHanaSource, SapOpenHubSource, SapTableSource, ServiceNowSource, ShopifySource, SparkSource, SqlDWSource, SqlMISource, SqlServerSource, SqlSource, SquareSource, SybaseSource, TeradataSource, VerticaSource, XeroSource, ZohoSource. All required parameters must be populated in order to send to Azure. @@ -691,7 +682,7 @@ class TabularSource(CopySource): } _subtype_map = { - 'type': {'AmazonMWSSource': 'AmazonMWSSource', 'AmazonRedshiftSource': 'AmazonRedshiftSource', 'AzureMariaDBSource': 'AzureMariaDBSource', 'AzureMySqlSource': 'AzureMySqlSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AzureSqlSource': 'AzureSqlSource', 'AzureTableSource': 'AzureTableSource', 'CassandraSource': 'CassandraSource', 'ConcurSource': 'ConcurSource', 'CouchbaseSource': 'CouchbaseSource', 'Db2Source': 'Db2Source', 'DrillSource': 'DrillSource', 'DynamicsAXSource': 'DynamicsAXSource', 'EloquaSource': 'EloquaSource', 'GoogleAdWordsSource': 'GoogleAdWordsSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'GreenplumSource': 'GreenplumSource', 'HBaseSource': 'HBaseSource', 'HiveSource': 'HiveSource', 'HubspotSource': 'HubspotSource', 'ImpalaSource': 'ImpalaSource', 'InformixSource': 'InformixSource', 'JiraSource': 'JiraSource', 'MagentoSource': 'MagentoSource', 'MariaDBSource': 'MariaDBSource', 'MarketoSource': 'MarketoSource', 'MySqlSource': 'MySqlSource', 'NetezzaSource': 'NetezzaSource', 'OdbcSource': 'OdbcSource', 'OracleServiceCloudSource': 'OracleServiceCloudSource', 'PaypalSource': 'PaypalSource', 'PhoenixSource': 'PhoenixSource', 'PostgreSqlSource': 'PostgreSqlSource', 'PrestoSource': 'PrestoSource', 'QuickBooksSource': 'QuickBooksSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'SalesforceSource': 'SalesforceSource', 'SapBwSource': 'SapBwSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SapEccSource': 'SapEccSource', 'SapHanaSource': 'SapHanaSource', 'SapOpenHubSource': 'SapOpenHubSource', 'SapTableSource': 'SapTableSource', 'ServiceNowSource': 'ServiceNowSource', 'ShopifySource': 'ShopifySource', 'SparkSource': 'SparkSource', 'SqlDWSource': 'SqlDWSource', 'SqlMISource': 'SqlMISource', 'SqlServerSource': 'SqlServerSource', 'SqlSource': 'SqlSource', 'SquareSource': 'SquareSource', 'SybaseSource': 'SybaseSource', 'TeradataSource': 'TeradataSource', 'VerticaSource': 'VerticaSource', 'XeroSource': 'XeroSource', 'ZohoSource': 'ZohoSource'} + 'type': {'AmazonMWSSource': 'AmazonMWSSource', 'AmazonRdsForSqlServerSource': 'AmazonRdsForSqlServerSource', 'AmazonRedshiftSource': 'AmazonRedshiftSource', 'AzureMariaDBSource': 'AzureMariaDBSource', 'AzureMySqlSource': 'AzureMySqlSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AzureSqlSource': 'AzureSqlSource', 'AzureTableSource': 'AzureTableSource', 'CassandraSource': 'CassandraSource', 'ConcurSource': 'ConcurSource', 'CouchbaseSource': 'CouchbaseSource', 'Db2Source': 'Db2Source', 'DrillSource': 'DrillSource', 'DynamicsAXSource': 'DynamicsAXSource', 'EloquaSource': 'EloquaSource', 'GoogleAdWordsSource': 'GoogleAdWordsSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'GreenplumSource': 'GreenplumSource', 'HBaseSource': 'HBaseSource', 'HiveSource': 'HiveSource', 'HubspotSource': 'HubspotSource', 'ImpalaSource': 'ImpalaSource', 'InformixSource': 'InformixSource', 'JiraSource': 'JiraSource', 'MagentoSource': 'MagentoSource', 'MariaDBSource': 'MariaDBSource', 'MarketoSource': 'MarketoSource', 'MySqlSource': 'MySqlSource', 'NetezzaSource': 'NetezzaSource', 'OdbcSource': 'OdbcSource', 'OracleServiceCloudSource': 'OracleServiceCloudSource', 'PaypalSource': 'PaypalSource', 'PhoenixSource': 'PhoenixSource', 'PostgreSqlSource': 'PostgreSqlSource', 'PrestoSource': 'PrestoSource', 'QuickBooksSource': 'QuickBooksSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'SalesforceSource': 'SalesforceSource', 'SapBwSource': 'SapBwSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SapEccSource': 'SapEccSource', 'SapHanaSource': 'SapHanaSource', 'SapOpenHubSource': 'SapOpenHubSource', 'SapTableSource': 'SapTableSource', 'ServiceNowSource': 'ServiceNowSource', 'ShopifySource': 'ShopifySource', 'SparkSource': 'SparkSource', 'SqlDWSource': 'SqlDWSource', 'SqlMISource': 'SqlMISource', 'SqlServerSource': 'SqlServerSource', 'SqlSource': 'SqlSource', 'SquareSource': 'SquareSource', 'SybaseSource': 'SybaseSource', 'TeradataSource': 'TeradataSource', 'VerticaSource': 'VerticaSource', 'XeroSource': 'XeroSource', 'ZohoSource': 'ZohoSource'} } def __init__( @@ -758,6 +749,433 @@ def __init__( self.query = kwargs.get('query', None) +class AmazonRdsForOracleLinkedService(LinkedService): + """AmazonRdsForOracle database. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are deserialized to this + collection. + :type additional_properties: dict[str, any] + :param type: Required. Type of linked service.Constant filled by server. + :type type: str + :param connect_via: The integration runtime reference. + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the linked service. + :type annotations: list[any] + :param connection_string: Required. The connection string. Type: string, SecureString or + AzureKeyVaultSecretReference. + :type connection_string: any + :param password: The Azure key vault secret reference of password in connection string. + :type password: ~azure.synapse.artifacts.models.SecretBase + :param encrypted_credential: The encrypted credential used for authentication. Credentials are + encrypted using the integration runtime credential manager. Type: string (or Expression with + resultType string). + :type encrypted_credential: any + """ + + _validation = { + 'type': {'required': True}, + 'connection_string': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__( + self, + **kwargs + ): + super(AmazonRdsForOracleLinkedService, self).__init__(**kwargs) + self.type = 'AmazonRdsForOracle' # type: str + self.connection_string = kwargs['connection_string'] + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) + + +class AmazonRdsForOraclePartitionSettings(msrest.serialization.Model): + """The settings that will be leveraged for AmazonRdsForOracle source partitioning. + + :param partition_names: Names of the physical partitions of AmazonRdsForOracle table. + :type partition_names: any + :param partition_column_name: The name of the column in integer type that will be used for + proceeding range partitioning. Type: string (or Expression with resultType string). + :type partition_column_name: any + :param partition_upper_bound: The maximum value of column specified in partitionColumnName that + will be used for proceeding range partitioning. Type: string (or Expression with resultType + string). + :type partition_upper_bound: any + :param partition_lower_bound: The minimum value of column specified in partitionColumnName that + will be used for proceeding range partitioning. Type: string (or Expression with resultType + string). + :type partition_lower_bound: any + """ + + _attribute_map = { + 'partition_names': {'key': 'partitionNames', 'type': 'object'}, + 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, + 'partition_upper_bound': {'key': 'partitionUpperBound', 'type': 'object'}, + 'partition_lower_bound': {'key': 'partitionLowerBound', 'type': 'object'}, + } + + def __init__( + self, + **kwargs + ): + super(AmazonRdsForOraclePartitionSettings, self).__init__(**kwargs) + self.partition_names = kwargs.get('partition_names', None) + self.partition_column_name = kwargs.get('partition_column_name', None) + self.partition_upper_bound = kwargs.get('partition_upper_bound', None) + self.partition_lower_bound = kwargs.get('partition_lower_bound', None) + + +class AmazonRdsForOracleSource(CopySource): + """A copy activity AmazonRdsForOracle source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are deserialized to this + collection. + :type additional_properties: dict[str, any] + :param type: Required. Copy source type.Constant filled by server. + :type type: str + :param source_retry_count: Source retry count. Type: integer (or Expression with resultType + integer). + :type source_retry_count: any + :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: any + :param max_concurrent_connections: The maximum concurrent connection count for the source data + store. Type: integer (or Expression with resultType integer). + :type max_concurrent_connections: any + :param oracle_reader_query: AmazonRdsForOracle reader query. Type: string (or Expression with + resultType string). + :type oracle_reader_query: any + :param query_timeout: Query timeout. Type: string (or Expression with resultType string), + pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type query_timeout: any + :param partition_option: The partition mechanism that will be used for AmazonRdsForOracle read + in parallel. Type: string (or Expression with resultType string). + :type partition_option: any + :param partition_settings: The settings that will be leveraged for AmazonRdsForOracle source + partitioning. + :type partition_settings: ~azure.synapse.artifacts.models.AmazonRdsForOraclePartitionSettings + :param additional_columns: Specifies the additional columns to be added to source data. Type: + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, + 'oracle_reader_query': {'key': 'oracleReaderQuery', 'type': 'object'}, + 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, + 'partition_settings': {'key': 'partitionSettings', 'type': 'AmazonRdsForOraclePartitionSettings'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, + } + + def __init__( + self, + **kwargs + ): + super(AmazonRdsForOracleSource, self).__init__(**kwargs) + self.type = 'AmazonRdsForOracleSource' # type: str + self.oracle_reader_query = kwargs.get('oracle_reader_query', None) + self.query_timeout = kwargs.get('query_timeout', None) + self.partition_option = kwargs.get('partition_option', None) + self.partition_settings = kwargs.get('partition_settings', None) + self.additional_columns = kwargs.get('additional_columns', None) + + +class AmazonRdsForOracleTableDataset(Dataset): + """The AmazonRdsForOracle database dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are deserialized to this + collection. + :type additional_properties: dict[str, any] + :param type: Required. Type of dataset.Constant filled by server. + :type type: str + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: array (or Expression + with resultType array), itemType: DatasetDataElement. + :type structure: any + :param schema: Columns that define the physical type schema of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetSchemaDataElement. + :type schema: any + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the Dataset. + :type annotations: list[any] + :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the + root level. + :type folder: ~azure.synapse.artifacts.models.DatasetFolder + :param schema_type_properties_schema: The schema name of the AmazonRdsForOracle database. Type: + string (or Expression with resultType string). + :type schema_type_properties_schema: any + :param table: The table name of the AmazonRdsForOracle database. Type: string (or Expression + with resultType string). + :type table: any + """ + + _validation = { + 'type': {'required': True}, + 'linked_service_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, + 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, + 'table': {'key': 'typeProperties.table', 'type': 'object'}, + } + + def __init__( + self, + **kwargs + ): + super(AmazonRdsForOracleTableDataset, self).__init__(**kwargs) + self.type = 'AmazonRdsForOracleTable' # type: str + self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) + self.table = kwargs.get('table', None) + + +class AmazonRdsForSqlServerLinkedService(LinkedService): + """Amazon RDS for SQL Server linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are deserialized to this + collection. + :type additional_properties: dict[str, any] + :param type: Required. Type of linked service.Constant filled by server. + :type type: str + :param connect_via: The integration runtime reference. + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the linked service. + :type annotations: list[any] + :param connection_string: Required. The connection string. Type: string, SecureString or + AzureKeyVaultSecretReference. + :type connection_string: any + :param user_name: The on-premises Windows authentication user name. Type: string (or Expression + with resultType string). + :type user_name: any + :param password: The on-premises Windows authentication password. + :type password: ~azure.synapse.artifacts.models.SecretBase + :param encrypted_credential: The encrypted credential used for authentication. Credentials are + encrypted using the integration runtime credential manager. Type: string (or Expression with + resultType string). + :type encrypted_credential: any + """ + + _validation = { + 'type': {'required': True}, + 'connection_string': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__( + self, + **kwargs + ): + super(AmazonRdsForSqlServerLinkedService, self).__init__(**kwargs) + self.type = 'AmazonRdsForSqlServer' # type: str + self.connection_string = kwargs['connection_string'] + self.user_name = kwargs.get('user_name', None) + self.password = kwargs.get('password', None) + self.encrypted_credential = kwargs.get('encrypted_credential', None) + + +class AmazonRdsForSqlServerSource(TabularSource): + """A copy activity Amazon RDS for SQL Server source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are deserialized to this + collection. + :type additional_properties: dict[str, any] + :param type: Required. Copy source type.Constant filled by server. + :type type: str + :param source_retry_count: Source retry count. Type: integer (or Expression with resultType + integer). + :type source_retry_count: any + :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: any + :param max_concurrent_connections: The maximum concurrent connection count for the source data + store. Type: integer (or Expression with resultType integer). + :type max_concurrent_connections: any + :param query_timeout: Query timeout. Type: string (or Expression with resultType string), + pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type query_timeout: any + :param additional_columns: Specifies the additional columns to be added to source data. Type: + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any + :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). + :type sql_reader_query: any + :param sql_reader_stored_procedure_name: Name of the stored procedure for a SQL Database + source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression + with resultType string). + :type sql_reader_stored_procedure_name: any + :param stored_procedure_parameters: Value and type setting for stored procedure parameters. + Example: "{Parameter1: {value: "1", type: "int"}}". + :type stored_procedure_parameters: dict[str, + ~azure.synapse.artifacts.models.StoredProcedureParameter] + :param produce_additional_types: Which additional types to produce. + :type produce_additional_types: any + :param partition_option: The partition mechanism that will be used for Sql read in parallel. + Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". + :type partition_option: any + :param partition_settings: The settings that will be leveraged for Sql source partitioning. + :type partition_settings: ~azure.synapse.artifacts.models.SqlPartitionSettings + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, + 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, + 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, + 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, + 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, + 'produce_additional_types': {'key': 'produceAdditionalTypes', 'type': 'object'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, + 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, + } + + def __init__( + self, + **kwargs + ): + super(AmazonRdsForSqlServerSource, self).__init__(**kwargs) + self.type = 'AmazonRdsForSqlServerSource' # type: str + self.sql_reader_query = kwargs.get('sql_reader_query', None) + self.sql_reader_stored_procedure_name = kwargs.get('sql_reader_stored_procedure_name', None) + self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) + self.produce_additional_types = kwargs.get('produce_additional_types', None) + self.partition_option = kwargs.get('partition_option', None) + self.partition_settings = kwargs.get('partition_settings', None) + + +class AmazonRdsForSqlServerTableDataset(Dataset): + """The Amazon RDS for SQL Server dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are deserialized to this + collection. + :type additional_properties: dict[str, any] + :param type: Required. Type of dataset.Constant filled by server. + :type type: str + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: array (or Expression + with resultType array), itemType: DatasetDataElement. + :type structure: any + :param schema: Columns that define the physical type schema of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetSchemaDataElement. + :type schema: any + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the Dataset. + :type annotations: list[any] + :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the + root level. + :type folder: ~azure.synapse.artifacts.models.DatasetFolder + :param schema_type_properties_schema: The schema name of the SQL Server dataset. Type: string + (or Expression with resultType string). + :type schema_type_properties_schema: any + :param table: The table name of the SQL Server dataset. Type: string (or Expression with + resultType string). + :type table: any + """ + + _validation = { + 'type': {'required': True}, + 'linked_service_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, + 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, + 'table': {'key': 'typeProperties.table', 'type': 'object'}, + } + + def __init__( + self, + **kwargs + ): + super(AmazonRdsForSqlServerTableDataset, self).__init__(**kwargs) + self.type = 'AmazonRdsForSqlServerTable' # type: str + self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) + self.table = kwargs.get('table', None) + + class AmazonRedshiftLinkedService(LinkedService): """Linked service for Amazon Redshift. @@ -769,13 +1187,11 @@ class AmazonRedshiftLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param server: Required. The name of the Amazon Redshift server. Type: string (or Expression @@ -785,7 +1201,7 @@ class AmazonRedshiftLinkedService(LinkedService): resultType string). :type username: any :param password: The password of the Amazon Redshift source. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param database: Required. The database name of the Amazon Redshift source. Type: string (or Expression with resultType string). :type database: any @@ -863,8 +1279,7 @@ class AmazonRedshiftSource(TabularSource): :param redshift_unload_settings: The Amazon S3 settings needed for the interim Amazon S3 when copying from Amazon Redshift with unload. With this, data from Amazon Redshift source will be unloaded into S3 first and then copied into the targeted sink from the interim S3. - :type redshift_unload_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.RedshiftUnloadSettings + :type redshift_unload_settings: ~azure.synapse.artifacts.models.RedshiftUnloadSettings """ _validation = { @@ -912,16 +1327,14 @@ class AmazonRedshiftTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -983,16 +1396,14 @@ class AmazonS3Dataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param bucket_name: Required. The name of the Amazon S3 bucket. Type: string (or Expression with resultType string). :type bucket_name: any @@ -1012,9 +1423,9 @@ class AmazonS3Dataset(Dataset): Expression with resultType string). :type modified_datetime_end: any :param format: The format of files. - :type format: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetStorageFormat + :type format: ~azure.synapse.artifacts.models.DatasetStorageFormat :param compression: The data compression method used for the Amazon S3 object. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression """ _validation = { @@ -1070,13 +1481,11 @@ class AmazonS3LinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param authentication_type: The authentication type of S3. Allowed value: AccessKey (default) @@ -1087,13 +1496,13 @@ class AmazonS3LinkedService(LinkedService): :type access_key_id: any :param secret_access_key: The secret access key of the Amazon S3 Identity and Access Management (IAM) user. - :type secret_access_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type secret_access_key: ~azure.synapse.artifacts.models.SecretBase :param service_url: This value specifies the endpoint to access with the S3 Connector. This is an optional property; change it only if you want to try a different service endpoint or want to switch between https and http. Type: string (or Expression with resultType string). :type service_url: any :param session_token: The session token for the S3 temporary security credential. - :type session_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type session_token: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -1370,9 +1779,9 @@ class ControlActivity(Activity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] """ _validation = { @@ -1416,9 +1825,9 @@ class AppendVariableActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param variable_name: Name of the variable whose value needs to be appended to. :type variable_name: str :param value: Value to be appended. Could be a static value or Expression. @@ -1544,18 +1953,16 @@ class AvroDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param location: The location of the avro storage. - :type location: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetLocation + :type location: ~azure.synapse.artifacts.models.DatasetLocation :param avro_compression_codec: A string from AvroCompressionCodecEnum or an expression. :type avro_compression_codec: any :param avro_compression_level: @@ -1762,9 +2169,9 @@ class AvroSink(CopySink): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Avro store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreWriteSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreWriteSettings :param format_settings: Avro format settings. - :type format_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.AvroWriteSettings + :type format_settings: ~azure.synapse.artifacts.models.AvroWriteSettings """ _validation = { @@ -1813,7 +2220,7 @@ class AvroSource(CopySource): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Avro store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :type additional_columns: any @@ -1939,20 +2346,18 @@ class AzureBatchLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param account_name: Required. The Azure Batch account name. Type: string (or Expression with resultType string). :type account_name: any :param access_key: The Azure Batch account access key. - :type access_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type access_key: ~azure.synapse.artifacts.models.SecretBase :param batch_uri: Required. The Azure Batch URI. Type: string (or Expression with resultType string). :type batch_uri: any @@ -1960,8 +2365,7 @@ class AzureBatchLinkedService(LinkedService): resultType string). :type pool_name: any :param linked_service_name: Required. The Azure Storage linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -2024,16 +2428,14 @@ class AzureBlobDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param folder_path: The path of the Azure Blob storage. Type: string (or Expression with resultType string). :type folder_path: any @@ -2050,9 +2452,9 @@ class AzureBlobDataset(Dataset): Expression with resultType string). :type modified_datetime_end: any :param format: The format of the Azure Blob storage. - :type format: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetStorageFormat + :type format: ~azure.synapse.artifacts.models.DatasetStorageFormat :param compression: The data compression method used for the blob storage. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression """ _validation = { @@ -2113,16 +2515,14 @@ class AzureBlobFSDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param folder_path: The path of the Azure Data Lake Storage Gen2 storage. Type: string (or Expression with resultType string). :type folder_path: any @@ -2130,9 +2530,9 @@ class AzureBlobFSDataset(Dataset): with resultType string). :type file_name: any :param format: The format of the Azure Data Lake Storage Gen2 storage. - :type format: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetStorageFormat + :type format: ~azure.synapse.artifacts.models.DatasetStorageFormat :param compression: The data compression method used for the blob storage. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression """ _validation = { @@ -2179,13 +2579,11 @@ class AzureBlobFSLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param url: Required. Endpoint for the Azure Data Lake Storage Gen2 service. Type: string (or @@ -2199,7 +2597,7 @@ class AzureBlobFSLinkedService(LinkedService): :type service_principal_id: any :param service_principal_key: The Key of the application used to authenticate against the Azure Data Lake Storage Gen2 account. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -2576,28 +2974,24 @@ class AzureBlobStorageLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: The connection string. It is mutually exclusive with sasUri, serviceEndpoint property. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param account_key: The Azure key vault secret reference of accountKey in connection string. - :type account_key: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type account_key: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param sas_uri: SAS URI of the Azure Blob Storage resource. It is mutually exclusive with connectionString, serviceEndpoint property. Type: string, SecureString or AzureKeyVaultSecretReference. :type sas_uri: any :param sas_token: The Azure key vault secret reference of sasToken in sas uri. - :type sas_token: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type sas_token: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param service_endpoint: Blob service endpoint of the Azure Blob Storage resource. It is mutually exclusive with connectionString, sasUri property. :type service_endpoint: str @@ -2606,7 +3000,7 @@ class AzureBlobStorageLinkedService(LinkedService): :type service_principal_id: any :param service_principal_key: The key of the service principal used to authenticate against Azure SQL Data Warehouse. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -2614,6 +3008,10 @@ class AzureBlobStorageLinkedService(LinkedService): values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data factory regions’ cloud type. Type: string (or Expression with resultType string). :type azure_cloud_type: any + :param account_kind: Specify the kind of your storage account. Allowed values are: Storage + (general purpose v1), StorageV2 (general purpose v2), BlobStorage, or BlockBlobStorage. Type: + string (or Expression with resultType string). + :type account_kind: str :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -2640,6 +3038,7 @@ class AzureBlobStorageLinkedService(LinkedService): 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, 'azure_cloud_type': {'key': 'typeProperties.azureCloudType', 'type': 'object'}, + 'account_kind': {'key': 'typeProperties.accountKind', 'type': 'str'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, } @@ -2658,6 +3057,7 @@ def __init__( self.service_principal_key = kwargs.get('service_principal_key', None) self.tenant = kwargs.get('tenant', None) self.azure_cloud_type = kwargs.get('azure_cloud_type', None) + self.account_kind = kwargs.get('account_kind', None) self.encrypted_credential = kwargs.get('encrypted_credential', None) @@ -2846,16 +3246,14 @@ class AzureDatabricksDeltaLakeDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table: The name of delta table. Type: string (or Expression with resultType string). :type table: any :param database: The database name of delta table. Type: string (or Expression with resultType @@ -3055,13 +3453,11 @@ class AzureDatabricksDeltaLakeLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param domain: Required. :code:``.azuredatabricks.net, domain name of your Databricks @@ -3070,7 +3466,7 @@ class AzureDatabricksDeltaLakeLinkedService(LinkedService): :param access_token: Required. Access token for databricks REST API. Refer to https://docs.azuredatabricks.net/api/latest/authentication.html. Type: string, SecureString or AzureKeyVaultSecretReference. - :type access_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type access_token: ~azure.synapse.artifacts.models.SecretBase :param cluster_id: The id of an existing interactive cluster that will be used for all runs of this job. Type: string (or Expression with resultType string). :type cluster_id: any @@ -3140,8 +3536,7 @@ class AzureDatabricksDeltaLakeSink(CopySink): string). :type pre_copy_script: any :param import_settings: Azure Databricks Delta Lake import settings. - :type import_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureDatabricksDeltaLakeImportCommand + :type import_settings: ~azure.synapse.artifacts.models.AzureDatabricksDeltaLakeImportCommand """ _validation = { @@ -3193,8 +3588,7 @@ class AzureDatabricksDeltaLakeSource(CopySource): resultType string). :type query: any :param export_settings: Azure Databricks Delta Lake export settings. - :type export_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureDatabricksDeltaLakeExportCommand + :type export_settings: ~azure.synapse.artifacts.models.AzureDatabricksDeltaLakeExportCommand """ _validation = { @@ -3232,13 +3626,11 @@ class AzureDatabricksLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param domain: Required. :code:``.azuredatabricks.net, domain name of your Databricks @@ -3247,7 +3639,7 @@ class AzureDatabricksLinkedService(LinkedService): :param access_token: Access token for databricks REST API. Refer to https://docs.azuredatabricks.net/api/latest/authentication.html. Type: string (or Expression with resultType string). - :type access_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type access_token: ~azure.synapse.artifacts.models.SecretBase :param authentication: Required to specify MSI, if using Workspace resource id for databricks REST API. Type: string (or Expression with resultType string). :type authentication: any @@ -3384,14 +3776,13 @@ class ExecutionActivity(Activity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy """ _validation = { @@ -3439,14 +3830,13 @@ class AzureDataExplorerCommandActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param command: Required. A control command, according to the Azure Data Explorer command syntax. Type: string (or Expression with resultType string). :type command: any @@ -3495,40 +3885,35 @@ class AzureDataExplorerLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param endpoint: Required. The endpoint of Azure Data Explorer (the engine's endpoint). URL will be in the format https://:code:``.:code:``.kusto.windows.net. Type: string (or Expression with resultType string). :type endpoint: any - :param service_principal_id: Required. The ID of the service principal used to authenticate - against Azure Data Explorer. Type: string (or Expression with resultType string). + :param service_principal_id: The ID of the service principal used to authenticate against Azure + Data Explorer. Type: string (or Expression with resultType string). :type service_principal_id: any - :param service_principal_key: Required. The key of the service principal used to authenticate - against Kusto. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :param service_principal_key: The key of the service principal used to authenticate against + Kusto. + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param database: Required. Database name for connection. Type: string (or Expression with resultType string). :type database: any - :param tenant: Required. The name or ID of the tenant to which the service principal belongs. - Type: string (or Expression with resultType string). + :param tenant: The name or ID of the tenant to which the service principal belongs. Type: + string (or Expression with resultType string). :type tenant: any """ _validation = { 'type': {'required': True}, 'endpoint': {'required': True}, - 'service_principal_id': {'required': True}, - 'service_principal_key': {'required': True}, 'database': {'required': True}, - 'tenant': {'required': True}, } _attribute_map = { @@ -3552,10 +3937,10 @@ def __init__( super(AzureDataExplorerLinkedService, self).__init__(**kwargs) self.type = 'AzureDataExplorer' # type: str self.endpoint = kwargs['endpoint'] - self.service_principal_id = kwargs['service_principal_id'] - self.service_principal_key = kwargs['service_principal_key'] + self.service_principal_id = kwargs.get('service_principal_id', None) + self.service_principal_key = kwargs.get('service_principal_key', None) self.database = kwargs['database'] - self.tenant = kwargs['tenant'] + self.tenant = kwargs.get('tenant', None) class AzureDataExplorerSink(CopySink): @@ -3703,16 +4088,14 @@ class AzureDataExplorerTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table: The table name of the Azure Data Explorer database. Type: string (or Expression with resultType string). :type table: any @@ -3756,13 +4139,11 @@ class AzureDataLakeAnalyticsLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param account_name: Required. The Azure Data Lake Analytics account name. Type: string (or @@ -3773,7 +4154,7 @@ class AzureDataLakeAnalyticsLinkedService(LinkedService): :type service_principal_id: any :param service_principal_key: The Key of the application used to authenticate against the Azure Data Lake Analytics account. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: Required. The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -3850,16 +4231,14 @@ class AzureDataLakeStoreDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param folder_path: Path to the folder in the Azure Data Lake Store. Type: string (or Expression with resultType string). :type folder_path: any @@ -3867,10 +4246,10 @@ class AzureDataLakeStoreDataset(Dataset): Expression with resultType string). :type file_name: any :param format: The format of the Data Lake Store. - :type format: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetStorageFormat + :type format: ~azure.synapse.artifacts.models.DatasetStorageFormat :param compression: The data compression method used for the item(s) in the Azure Data Lake Store. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression """ _validation = { @@ -3917,13 +4296,11 @@ class AzureDataLakeStoreLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param data_lake_store_uri: Required. Data Lake Store service URI. Type: string (or Expression @@ -3934,7 +4311,7 @@ class AzureDataLakeStoreLinkedService(LinkedService): :type service_principal_id: any :param service_principal_key: The Key of the application used to authenticate against the Azure Data Lake Store account. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -4356,13 +4733,11 @@ class AzureFileStorageLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. Host name of the server. Type: string (or Expression with resultType @@ -4372,19 +4747,17 @@ class AzureFileStorageLinkedService(LinkedService): string). :type user_id: any :param password: Password to logon the server. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param connection_string: The connection string. It is mutually exclusive with sasUri property. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param account_key: The Azure key vault secret reference of accountKey in connection string. - :type account_key: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type account_key: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param sas_uri: SAS URI of the Azure File resource. It is mutually exclusive with connectionString property. Type: string, SecureString or AzureKeyVaultSecretReference. :type sas_uri: any :param sas_token: The Azure key vault secret reference of sasToken in sas uri. - :type sas_token: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type sas_token: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param file_share: The azure file share name. It is required when auth with accountKey/sasToken. Type: string (or Expression with resultType string). :type file_share: any @@ -4610,18 +4983,16 @@ class AzureFunctionActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param method: Required. Rest API method for target endpoint. Possible values include: "GET", "POST", "PUT", "DELETE", "OPTIONS", "HEAD", "TRACE". - :type method: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureFunctionActivityMethod + :type method: str or ~azure.synapse.artifacts.models.AzureFunctionActivityMethod :param function_name: Required. Name of the Function that the Azure Function Activity will call. Type: string (or Expression with resultType string). :type function_name: any @@ -4679,20 +5050,18 @@ class AzureFunctionLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param function_app_url: Required. The endpoint of the Azure Function App. URL will be in the format https://:code:``.azurewebsites.net. :type function_app_url: any :param function_key: Function or Host key for Azure Function App. - :type function_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type function_key: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4738,13 +5107,11 @@ class AzureKeyVaultLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param base_url: Required. The base URL of the Azure Key Vault. e.g. @@ -4816,7 +5183,7 @@ class AzureKeyVaultSecretReference(SecretBase): :param type: Required. Type of the secret.Constant filled by server. :type type: str :param store: Required. The Azure Key Vault linked service reference. - :type store: ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type store: ~azure.synapse.artifacts.models.LinkedServiceReference :param secret_name: Required. The name of the secret in Azure Key Vault. Type: string (or Expression with resultType string). :type secret_name: any @@ -4860,20 +5227,18 @@ class AzureMariaDBLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type pwd: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4980,16 +5345,14 @@ class AzureMariaDBTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -5036,14 +5399,13 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param global_parameters: Key,Value pairs to be passed to the Azure ML Batch Execution Service endpoint. Keys must match the names of web service parameters defined in the published Azure ML web service. Values will be passed in the GlobalParameters property of the Azure ML batch @@ -5053,14 +5415,12 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): Service Outputs to AzureMLWebServiceFile objects specifying the output Blob locations. This information will be passed in the WebServiceOutputs property of the Azure ML batch execution request. - :type web_service_outputs: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureMLWebServiceFile] + :type web_service_outputs: dict[str, ~azure.synapse.artifacts.models.AzureMLWebServiceFile] :param web_service_inputs: Key,Value pairs, mapping the names of Azure ML endpoint's Web Service Inputs to AzureMLWebServiceFile objects specifying the input Blob locations.. This information will be passed in the WebServiceInputs property of the Azure ML batch execution request. - :type web_service_inputs: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureMLWebServiceFile] + :type web_service_inputs: dict[str, ~azure.synapse.artifacts.models.AzureMLWebServiceFile] """ _validation = { @@ -5108,14 +5468,13 @@ class AzureMLExecutePipelineActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param ml_pipeline_id: Required. ID of the published Azure ML pipeline. Type: string (or Expression with resultType string). :type ml_pipeline_id: any @@ -5185,20 +5544,18 @@ class AzureMLLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param ml_endpoint: Required. The Batch Execution REST URL for an Azure ML Studio Web Service endpoint. Type: string (or Expression with resultType string). :type ml_endpoint: any :param api_key: Required. The API key for accessing the Azure ML model endpoint. - :type api_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type api_key: ~azure.synapse.artifacts.models.SecretBase :param update_resource_endpoint: The Update Resource REST URL for an Azure ML Studio Web Service endpoint. Type: string (or Expression with resultType string). :type update_resource_endpoint: any @@ -5208,7 +5565,7 @@ class AzureMLLinkedService(LinkedService): :type service_principal_id: any :param service_principal_key: The key of the service principal used to authenticate against the ARM-based updateResourceEndpoint of an Azure ML Studio web service. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -5266,13 +5623,11 @@ class AzureMLServiceLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param subscription_id: Required. Azure ML Service workspace subscription ID. Type: string (or @@ -5290,7 +5645,7 @@ class AzureMLServiceLinkedService(LinkedService): :type service_principal_id: any :param service_principal_key: The key of the service principal used to authenticate against the endpoint of a published Azure ML Service pipeline. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -5353,21 +5708,19 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param trained_model_name: Required. Name of the Trained Model module in the Web Service experiment to be updated. Type: string (or Expression with resultType string). :type trained_model_name: any :param trained_model_linked_service_name: Required. Name of Azure Storage linked service holding the .ilearner file that will be uploaded by the update operation. - :type trained_model_linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type trained_model_linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param trained_model_file_path: Required. The relative file path in trainedModelLinkedService to represent the .ilearner file that will be uploaded by the update operation. Type: string (or Expression with resultType string). @@ -5417,8 +5770,7 @@ class AzureMLWebServiceFile(msrest.serialization.Model): :type file_path: any :param linked_service_name: Required. Reference to an Azure Storage LinkedService, where Azure ML WebService Input/Output file located. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference """ _validation = { @@ -5451,21 +5803,18 @@ class AzureMySqlLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param password: The Azure key vault secret reference of password in connection string. - :type password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type password: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5626,16 +5975,14 @@ class AzureMySqlTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The Azure MySQL database table name. Type: string (or Expression with resultType string). :type table_name: any @@ -5684,21 +6031,18 @@ class AzurePostgreSqlLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param password: The Azure key vault secret reference of password in connection string. - :type password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type password: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5859,16 +6203,14 @@ class AzurePostgreSqlTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name of the Azure PostgreSQL database which includes both schema and table. Type: string (or Expression with resultType string). :type table_name: any @@ -5979,16 +6321,14 @@ class AzureSearchIndexDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param index_name: Required. The name of the Azure Search Index. Type: string (or Expression with resultType string). :type index_name: any @@ -6049,8 +6389,7 @@ class AzureSearchIndexSink(CopySink): :type max_concurrent_connections: any :param write_behavior: Specify the write behavior when upserting documents into Azure Search Index. Possible values include: "Merge", "Upload". - :type write_behavior: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureSearchIndexWriteBehaviorType + :type write_behavior: str or ~azure.synapse.artifacts.models.AzureSearchIndexWriteBehaviorType """ _validation = { @@ -6088,20 +6427,18 @@ class AzureSearchLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param url: Required. URL for Azure Search service. Type: string (or Expression with resultType string). :type url: any :param key: Admin Key for Azure Search service. - :type key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type key: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6147,27 +6484,24 @@ class AzureSqlDatabaseLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param password: The Azure key vault secret reference of password in connection string. - :type password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type password: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param service_principal_id: The ID of the service principal used to authenticate against Azure SQL Database. Type: string (or Expression with resultType string). :type service_principal_id: any :param service_principal_key: The key of the service principal used to authenticate against Azure SQL Database. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -6228,27 +6562,24 @@ class AzureSqlDWLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param password: The Azure key vault secret reference of password in connection string. - :type password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type password: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param service_principal_id: The ID of the service principal used to authenticate against Azure SQL Data Warehouse. Type: string (or Expression with resultType string). :type service_principal_id: any :param service_principal_key: The key of the service principal used to authenticate against Azure SQL Data Warehouse. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -6317,16 +6648,14 @@ class AzureSqlDWTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -6380,27 +6709,24 @@ class AzureSqlMILinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param password: The Azure key vault secret reference of password in connection string. - :type password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type password: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param service_principal_id: The ID of the service principal used to authenticate against Azure SQL Managed Instance. Type: string (or Expression with resultType string). :type service_principal_id: any :param service_principal_key: The key of the service principal used to authenticate against Azure SQL Managed Instance. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -6469,16 +6795,14 @@ class AzureSqlMITableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -6557,7 +6881,7 @@ class AzureSqlSink(CopySink): :type pre_copy_script: any :param stored_procedure_parameters: SQL stored procedure parameters. :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] :param stored_procedure_table_type_parameter_name: The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). :type stored_procedure_table_type_parameter_name: any @@ -6634,15 +6958,14 @@ class AzureSqlSource(TabularSource): :param stored_procedure_parameters: Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] :param produce_additional_types: Which additional types to produce. :type produce_additional_types: any :param partition_option: The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". :type partition_option: any :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.SqlPartitionSettings """ _validation = { @@ -6698,16 +7021,14 @@ class AzureSqlTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -6761,27 +7082,23 @@ class AzureStorageLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: The connection string. It is mutually exclusive with sasUri property. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param account_key: The Azure key vault secret reference of accountKey in connection string. - :type account_key: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type account_key: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param sas_uri: SAS URI of the Azure Storage resource. It is mutually exclusive with connectionString property. Type: string, SecureString or AzureKeyVaultSecretReference. :type sas_uri: any :param sas_token: The Azure key vault secret reference of sasToken in sas uri. - :type sas_token: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type sas_token: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6838,16 +7155,14 @@ class AzureTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: Required. The table name of the Azure Table storage. Type: string (or Expression with resultType string). :type table_name: any @@ -7020,27 +7335,23 @@ class AzureTableStorageLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: The connection string. It is mutually exclusive with sasUri property. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param account_key: The Azure key vault secret reference of accountKey in connection string. - :type account_key: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type account_key: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param sas_uri: SAS URI of the Azure Storage resource. It is mutually exclusive with connectionString property. Type: string, SecureString or AzureKeyVaultSecretReference. :type sas_uri: any :param sas_token: The Azure key vault secret reference of sasToken in sas uri. - :type sas_token: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type sas_token: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -7085,7 +7396,7 @@ class BigDataPoolReference(msrest.serialization.Model): :param type: Required. Big data pool reference type. Possible values include: "BigDataPoolReference". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.BigDataPoolReferenceType + :type type: str or ~azure.synapse.artifacts.models.BigDataPoolReferenceType :param reference_name: Required. Reference big data pool name. :type reference_name: str """ @@ -7176,11 +7487,11 @@ class BigDataPoolResourceInfo(TrackedResource): :param provisioning_state: The state of the Big Data pool. :type provisioning_state: str :param auto_scale: Auto-scaling properties. - :type auto_scale: ~azure.synapse.artifacts.v2021_06_01_preview.models.AutoScaleProperties + :type auto_scale: ~azure.synapse.artifacts.models.AutoScaleProperties :param creation_date: The time when the Big Data pool was created. :type creation_date: ~datetime.datetime :param auto_pause: Auto-pausing properties. - :type auto_pause: ~azure.synapse.artifacts.v2021_06_01_preview.models.AutoPauseProperties + :type auto_pause: ~azure.synapse.artifacts.models.AutoPauseProperties :param is_compute_isolation_enabled: Whether compute isolation is required or not. :type is_compute_isolation_enabled: bool :param session_level_packages_enabled: Whether session level packages enabled. @@ -7188,31 +7499,27 @@ class BigDataPoolResourceInfo(TrackedResource): :param cache_size: The cache size. :type cache_size: int :param dynamic_executor_allocation: Dynamic Executor Allocation. - :type dynamic_executor_allocation: - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicExecutorAllocation + :type dynamic_executor_allocation: ~azure.synapse.artifacts.models.DynamicExecutorAllocation :param spark_events_folder: The Spark events folder. :type spark_events_folder: str :param node_count: The number of nodes in the Big Data pool. :type node_count: int :param library_requirements: Library version requirements. - :type library_requirements: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LibraryRequirements + :type library_requirements: ~azure.synapse.artifacts.models.LibraryRequirements :param custom_libraries: List of custom libraries/packages associated with the spark pool. - :type custom_libraries: list[~azure.synapse.artifacts.v2021_06_01_preview.models.LibraryInfo] + :type custom_libraries: list[~azure.synapse.artifacts.models.LibraryInfo] :param spark_config_properties: Spark configuration file to specify additional properties. - :type spark_config_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LibraryRequirements + :type spark_config_properties: ~azure.synapse.artifacts.models.LibraryRequirements :param spark_version: The Apache Spark version. :type spark_version: str :param default_spark_log_folder: The default folder where Spark logs will be written. :type default_spark_log_folder: str :param node_size: The level of compute power that each node in the Big Data pool has. Possible values include: "None", "Small", "Medium", "Large", "XLarge", "XXLarge", "XXXLarge". - :type node_size: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.NodeSize + :type node_size: str or ~azure.synapse.artifacts.models.NodeSize :param node_size_family: The kind of nodes that the Big Data pool provides. Possible values include: "None", "MemoryOptimized". - :type node_size_family: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.NodeSizeFamily + :type node_size_family: str or ~azure.synapse.artifacts.models.NodeSizeFamily :ivar last_succeeded_timestamp: The time when the Big Data pool was updated successfully. :vartype last_succeeded_timestamp: ~datetime.datetime """ @@ -7282,7 +7589,7 @@ class BigDataPoolResourceInfoListResult(msrest.serialization.Model): :param next_link: Link to the next page of results. :type next_link: str :param value: List of Big Data pools. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.BigDataPoolResourceInfo] + :type value: list[~azure.synapse.artifacts.models.BigDataPoolResourceInfo] """ _attribute_map = { @@ -7318,20 +7625,18 @@ class BinaryDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param location: The location of the Binary storage. - :type location: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetLocation + :type location: ~azure.synapse.artifacts.models.DatasetLocation :param compression: The data compression method used for the binary dataset. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression """ _validation = { @@ -7411,8 +7716,7 @@ class BinaryReadSettings(FormatReadSettings): :param type: Required. The read setting type.Constant filled by server. :type type: str :param compression_properties: Compression settings. - :type compression_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.CompressionReadSettings + :type compression_properties: ~azure.synapse.artifacts.models.CompressionReadSettings """ _validation = { @@ -7460,7 +7764,7 @@ class BinarySink(CopySink): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Binary store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreWriteSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreWriteSettings """ _validation = { @@ -7507,9 +7811,9 @@ class BinarySource(CopySource): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Binary store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings :param format_settings: Binary format settings. - :type format_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.BinaryReadSettings + :type format_settings: ~azure.synapse.artifacts.models.BinaryReadSettings """ _validation = { @@ -7555,8 +7859,7 @@ class Trigger(msrest.serialization.Model): :type description: str :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRuntimeState + :vartype runtime_state: str or ~azure.synapse.artifacts.models.TriggerRuntimeState :param annotations: List of tags that can be used for describing the trigger. :type annotations: list[any] """ @@ -7609,13 +7912,11 @@ class MultiplePipelineTrigger(Trigger): :type description: str :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRuntimeState + :vartype runtime_state: str or ~azure.synapse.artifacts.models.TriggerRuntimeState :param annotations: List of tags that can be used for describing the trigger. :type annotations: list[any] :param pipelines: Pipelines that need to be started. - :type pipelines: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerPipelineReference] + :type pipelines: list[~azure.synapse.artifacts.models.TriggerPipelineReference] """ _validation = { @@ -7661,13 +7962,11 @@ class BlobEventsTrigger(MultiplePipelineTrigger): :type description: str :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRuntimeState + :vartype runtime_state: str or ~azure.synapse.artifacts.models.TriggerRuntimeState :param annotations: List of tags that can be used for describing the trigger. :type annotations: list[any] :param pipelines: Pipelines that need to be started. - :type pipelines: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerPipelineReference] + :type pipelines: list[~azure.synapse.artifacts.models.TriggerPipelineReference] :param blob_path_begins_with: The blob path must begin with the pattern provided for trigger to fire. For example, '/records/blobs/december/' will only fire the trigger for blobs in the december folder under the records container. At least one of these must be provided: @@ -7680,7 +7979,7 @@ class BlobEventsTrigger(MultiplePipelineTrigger): :param ignore_empty_blobs: If set to true, blobs with zero bytes will be ignored. :type ignore_empty_blobs: bool :param events: Required. The type of events that cause this trigger to fire. - :type events: list[str or ~azure.synapse.artifacts.v2021_06_01_preview.models.BlobEventType] + :type events: list[str or ~azure.synapse.artifacts.models.BlobEventType] :param scope: Required. The ARM resource ID of the Storage Account. :type scope: str """ @@ -7859,21 +8158,18 @@ class BlobTrigger(MultiplePipelineTrigger): :type description: str :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRuntimeState + :vartype runtime_state: str or ~azure.synapse.artifacts.models.TriggerRuntimeState :param annotations: List of tags that can be used for describing the trigger. :type annotations: list[any] :param pipelines: Pipelines that need to be started. - :type pipelines: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerPipelineReference] + :type pipelines: list[~azure.synapse.artifacts.models.TriggerPipelineReference] :param folder_path: Required. The path of the container/folder that will trigger the pipeline. :type folder_path: str :param max_concurrency: Required. The max number of parallel files to handle when it is triggered. :type max_concurrency: int :param linked_service: Required. The Azure Storage linked service reference. - :type linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference """ _validation = { @@ -7918,13 +8214,11 @@ class CassandraLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. Host name for connection. Type: string (or Expression with resultType @@ -7940,7 +8234,7 @@ class CassandraLinkedService(LinkedService): string). :type username: any :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -8016,7 +8310,7 @@ class CassandraSource(TabularSource): Possible values include: "ALL", "EACH_QUORUM", "QUORUM", "LOCAL_QUORUM", "ONE", "TWO", "THREE", "LOCAL_ONE", "SERIAL", "LOCAL_SERIAL". :type consistency_level: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.CassandraSourceReadConsistencyLevels + ~azure.synapse.artifacts.models.CassandraSourceReadConsistencyLevels """ _validation = { @@ -8064,16 +8358,14 @@ class CassandraTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name of the Cassandra database. Type: string (or Expression with resultType string). :type table_name: any @@ -8127,15 +8419,14 @@ class ChainingTrigger(Trigger): :type description: str :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRuntimeState + :vartype runtime_state: str or ~azure.synapse.artifacts.models.TriggerRuntimeState :param annotations: List of tags that can be used for describing the trigger. :type annotations: list[any] :param pipeline: Required. Pipeline for which runs are created when all upstream pipelines complete successfully. - :type pipeline: ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerPipelineReference + :type pipeline: ~azure.synapse.artifacts.models.TriggerPipelineReference :param depends_on: Required. Upstream Pipelines. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineReference] + :type depends_on: list[~azure.synapse.artifacts.models.PipelineReference] :param run_dimension: Required. Run Dimension property that needs to be emitted by upstream pipelines. :type run_dimension: str @@ -8183,7 +8474,7 @@ class CloudError(msrest.serialization.Model): :param target: Property name/path in request associated with error. :type target: str :param details: Array with additional error details. - :type details: list[~azure.synapse.artifacts.v2021_06_01_preview.models.CloudError] + :type details: list[~azure.synapse.artifacts.models.CloudError] """ _validation = { @@ -8209,6 +8500,44 @@ def __init__( self.details = kwargs.get('details', None) +class CloudErrorAutoGenerated(msrest.serialization.Model): + """The object that defines the structure of an Azure Synapse error response. + + All required parameters must be populated in order to send to Azure. + + :param code: Required. Error code. + :type code: str + :param message: Required. Error message. + :type message: str + :param target: Property name/path in request associated with error. + :type target: str + :param details: Array with additional error details. + :type details: list[~azure.synapse.artifacts.models.CloudErrorAutoGenerated] + """ + + _validation = { + 'code': {'required': True}, + 'message': {'required': True}, + } + + _attribute_map = { + 'code': {'key': 'error.code', 'type': 'str'}, + 'message': {'key': 'error.message', 'type': 'str'}, + 'target': {'key': 'error.target', 'type': 'str'}, + 'details': {'key': 'error.details', 'type': '[CloudErrorAutoGenerated]'}, + } + + def __init__( + self, + **kwargs + ): + super(CloudErrorAutoGenerated, self).__init__(**kwargs) + self.code = kwargs['code'] + self.message = kwargs['message'] + self.target = kwargs.get('target', None) + self.details = kwargs.get('details', None) + + class CommonDataServiceForAppsEntityDataset(Dataset): """The Common Data Service for Apps entity dataset. @@ -8228,16 +8557,14 @@ class CommonDataServiceForAppsEntityDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param entity_name: The logical name of the entity. Type: string (or Expression with resultType string). :type entity_name: any @@ -8281,21 +8608,18 @@ class CommonDataServiceForAppsLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param deployment_type: Required. The deployment type of the Common Data Service for Apps instance. 'Online' for Common Data Service for Apps Online and 'OnPremisesWithIfd' for Common Data Service for Apps on-premises with Ifd. Type: string (or Expression with resultType - string). Possible values include: "Online", "OnPremisesWithIfd". - :type deployment_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsDeploymentType + string). + :type deployment_type: any :param host_name: The host name of the on-premises Common Data Service for Apps server. The property is required for on-prem and not allowed for online. Type: string (or Expression with resultType string). @@ -8316,15 +8640,13 @@ class CommonDataServiceForAppsLinkedService(LinkedService): :param authentication_type: Required. The authentication type to connect to Common Data Service for Apps server. 'Office365' for online scenario, 'Ifd' for on-premises with Ifd scenario. 'AADServicePrincipal' for Server-To-Server authentication in online scenario. Type: string (or - Expression with resultType string). Possible values include: "Office365", "Ifd", - "AADServicePrincipal". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsAuthenticationType + Expression with resultType string). + :type authentication_type: any :param username: User name to access the Common Data Service for Apps instance. Type: string (or Expression with resultType string). :type username: any :param password: Password to access the Common Data Service for Apps instance. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param service_principal_id: The client ID of the application in Azure Active Directory used for Server-To-Server authentication. Type: string (or Expression with resultType string). :type service_principal_id: any @@ -8336,8 +8658,7 @@ class CommonDataServiceForAppsLinkedService(LinkedService): servicePrincipalCredential can be SecureString or AzureKeyVaultSecretReference. If servicePrincipalCredentialType is 'ServicePrincipalCert', servicePrincipalCredential can only be AzureKeyVaultSecretReference. - :type service_principal_credential: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_credential: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -8357,12 +8678,12 @@ class CommonDataServiceForAppsLinkedService(LinkedService): 'description': {'key': 'description', 'type': 'str'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'str'}, + 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'object'}, 'host_name': {'key': 'typeProperties.hostName', 'type': 'object'}, 'port': {'key': 'typeProperties.port', 'type': 'object'}, 'service_uri': {'key': 'typeProperties.serviceUri', 'type': 'object'}, 'organization_name': {'key': 'typeProperties.organizationName', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, 'username': {'key': 'typeProperties.username', 'type': 'object'}, 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, @@ -8418,8 +8739,7 @@ class CommonDataServiceForAppsSink(CopySink): :type max_concurrent_connections: any :param write_behavior: Required. The write behavior for the operation. Possible values include: "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsSinkWriteBehavior + :type write_behavior: str or ~azure.synapse.artifacts.models.DynamicsSinkWriteBehavior :param ignore_null_values: The flag indicating whether to ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). @@ -8557,13 +8877,11 @@ class ConcurLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_properties: Properties used to connect to Concur. It is mutually exclusive @@ -8575,7 +8893,7 @@ class ConcurLinkedService(LinkedService): :type username: any :param password: The password corresponding to the user name that you provided in the username field. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -8650,16 +8968,14 @@ class ConcurObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -8760,29 +9076,28 @@ class CopyActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param inputs: List of inputs for the activity. - :type inputs: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference] + :type inputs: list[~azure.synapse.artifacts.models.DatasetReference] :param outputs: List of outputs for the activity. - :type outputs: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference] + :type outputs: list[~azure.synapse.artifacts.models.DatasetReference] :param source: Required. Copy activity source. - :type source: ~azure.synapse.artifacts.v2021_06_01_preview.models.CopySource + :type source: ~azure.synapse.artifacts.models.CopySource :param sink: Required. Copy activity sink. - :type sink: ~azure.synapse.artifacts.v2021_06_01_preview.models.CopySink + :type sink: ~azure.synapse.artifacts.models.CopySink :param translator: Copy activity translator. If not specified, tabular translator is used. :type translator: any :param enable_staging: Specifies whether to copy data via an interim staging. Default value is false. Type: boolean (or Expression with resultType boolean). :type enable_staging: any :param staging_settings: Specifies interim staging settings when EnableStaging is true. - :type staging_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StagingSettings + :type staging_settings: ~azure.synapse.artifacts.models.StagingSettings :param parallel_copies: Maximum number of concurrent sessions opened on the source or sink to avoid overloading the data store. Type: integer (or Expression with resultType integer), minimum: 0. @@ -8796,13 +9111,12 @@ class CopyActivity(ExecutionActivity): :param redirect_incompatible_row_settings: Redirect incompatible row settings when EnableSkipIncompatibleRow is true. :type redirect_incompatible_row_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.RedirectIncompatibleRowSettings + ~azure.synapse.artifacts.models.RedirectIncompatibleRowSettings :param log_storage_settings: (Deprecated. Please use LogSettings) Log storage settings customer need to provide when enabling session log. - :type log_storage_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LogStorageSettings + :type log_storage_settings: ~azure.synapse.artifacts.models.LogStorageSettings :param log_settings: Log settings customer needs provide when enabling log. - :type log_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.LogSettings + :type log_settings: ~azure.synapse.artifacts.models.LogSettings :param preserve_rules: Preserve Rules. :type preserve_rules: list[any] :param preserve: Preserve rules. @@ -8811,7 +9125,7 @@ class CopyActivity(ExecutionActivity): (or Expression with resultType boolean). :type validate_data_consistency: any :param skip_error_file: Specify the fault tolerance for data consistency. - :type skip_error_file: ~azure.synapse.artifacts.v2021_06_01_preview.models.SkipErrorFile + :type skip_error_file: ~azure.synapse.artifacts.models.SkipErrorFile """ _validation = { @@ -8947,13 +9261,11 @@ class CosmosDbLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: The connection string. Type: string, SecureString or @@ -8966,7 +9278,7 @@ class CosmosDbLinkedService(LinkedService): :type database: any :param account_key: The account key of the Azure CosmosDB account. Type: SecureString or AzureKeyVaultSecretReference. - :type account_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type account_key: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -9023,16 +9335,14 @@ class CosmosDbMongoDbApiCollectionDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param collection: Required. The collection name of the CosmosDB (MongoDB API) database. Type: string (or Expression with resultType string). :type collection: any @@ -9077,13 +9387,11 @@ class CosmosDbMongoDbApiLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The CosmosDB (MongoDB API) connection string. Type: string, @@ -9201,8 +9509,7 @@ class CosmosDbMongoDbApiSource(CopySource): with resultType string). :type filter: any :param cursor_methods: Cursor methods for Mongodb query. - :type cursor_methods: - ~azure.synapse.artifacts.v2021_06_01_preview.models.MongoDbCursorMethodsProperties + :type cursor_methods: ~azure.synapse.artifacts.models.MongoDbCursorMethodsProperties :param batch_size: Specifies the number of documents to return in each batch of the response from MongoDB instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. @@ -9265,16 +9572,14 @@ class CosmosDbSqlApiCollectionDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param collection_name: Required. CosmosDB (SQL API) collection name. Type: string (or Expression with resultType string). :type collection_name: any @@ -9438,21 +9743,18 @@ class CouchbaseLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param cred_string: The Azure key vault secret reference of credString in connection string. - :type cred_string: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type cred_string: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -9559,16 +9861,14 @@ class CouchbaseTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -9603,28 +9903,23 @@ def __init__( class CreateDataFlowDebugSessionRequest(msrest.serialization.Model): """Request body structure for creating data flow debug session. - :param data_flow_name: The name of the data flow. - :type data_flow_name: str - :param existing_cluster_id: The ID of existing Databricks cluster. - :type existing_cluster_id: str - :param cluster_timeout: Timeout setting for Databricks cluster. - :type cluster_timeout: int - :param new_cluster_name: The name of new Databricks cluster. - :type new_cluster_name: str - :param new_cluster_node_type: The type of new Databricks cluster. - :type new_cluster_node_type: str - :param data_bricks_linked_service: Data bricks linked service. - :type data_bricks_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceResource + :param compute_type: Compute type of the cluster. The value will be overwritten by the same + setting in integration runtime if provided. + :type compute_type: str + :param core_count: Core count of the cluster. The value will be overwritten by the same setting + in integration runtime if provided. + :type core_count: int + :param time_to_live: Time to live setting of the cluster in minutes. + :type time_to_live: int + :param integration_runtime: Set to use integration runtime setting for data flow debug session. + :type integration_runtime: ~azure.synapse.artifacts.models.IntegrationRuntimeDebugResource """ _attribute_map = { - 'data_flow_name': {'key': 'dataFlowName', 'type': 'str'}, - 'existing_cluster_id': {'key': 'existingClusterId', 'type': 'str'}, - 'cluster_timeout': {'key': 'clusterTimeout', 'type': 'int'}, - 'new_cluster_name': {'key': 'newClusterName', 'type': 'str'}, - 'new_cluster_node_type': {'key': 'newClusterNodeType', 'type': 'str'}, - 'data_bricks_linked_service': {'key': 'dataBricksLinkedService', 'type': 'LinkedServiceResource'}, + 'compute_type': {'key': 'computeType', 'type': 'str'}, + 'core_count': {'key': 'coreCount', 'type': 'int'}, + 'time_to_live': {'key': 'timeToLive', 'type': 'int'}, + 'integration_runtime': {'key': 'integrationRuntime', 'type': 'IntegrationRuntimeDebugResource'}, } def __init__( @@ -9632,12 +9927,10 @@ def __init__( **kwargs ): super(CreateDataFlowDebugSessionRequest, self).__init__(**kwargs) - self.data_flow_name = kwargs.get('data_flow_name', None) - self.existing_cluster_id = kwargs.get('existing_cluster_id', None) - self.cluster_timeout = kwargs.get('cluster_timeout', None) - self.new_cluster_name = kwargs.get('new_cluster_name', None) - self.new_cluster_node_type = kwargs.get('new_cluster_node_type', None) - self.data_bricks_linked_service = kwargs.get('data_bricks_linked_service', None) + self.compute_type = kwargs.get('compute_type', None) + self.core_count = kwargs.get('core_count', None) + self.time_to_live = kwargs.get('time_to_live', None) + self.integration_runtime = kwargs.get('integration_runtime', None) class CreateDataFlowDebugSessionResponse(msrest.serialization.Model): @@ -9699,26 +9992,23 @@ class CustomActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param command: Required. Command for custom activity Type: string (or Expression with resultType string). :type command: any :param resource_linked_service: Resource linked service reference. - :type resource_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type resource_linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param folder_path: Folder path for resource files Type: string (or Expression with resultType string). :type folder_path: any :param reference_objects: Reference objects. - :type reference_objects: - ~azure.synapse.artifacts.v2021_06_01_preview.models.CustomActivityReferenceObject + :type reference_objects: ~azure.synapse.artifacts.models.CustomActivityReferenceObject :param extended_properties: User defined property bag. There is no restriction on the keys or values that can be used. The user specified custom activity has the full responsibility to consume and interpret the content defined. @@ -9774,10 +10064,9 @@ class CustomActivityReferenceObject(msrest.serialization.Model): """Reference objects for custom activity. :param linked_services: Linked service references. - :type linked_services: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference] + :type linked_services: list[~azure.synapse.artifacts.models.LinkedServiceReference] :param datasets: Dataset references. - :type datasets: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference] + :type datasets: list[~azure.synapse.artifacts.models.DatasetReference] """ _attribute_map = { @@ -9813,16 +10102,14 @@ class CustomDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param type_properties: Custom dataset properties. :type type_properties: any """ @@ -9865,13 +10152,11 @@ class CustomDataSourceLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param type_properties: Required. Custom linked service properties. @@ -9910,7 +10195,7 @@ class CustomerManagedKeyDetails(msrest.serialization.Model): :ivar status: The customer managed key status on the workspace. :vartype status: str :param key: The key object of the workspace. - :type key: ~azure.synapse.artifacts.v2021_06_01_preview.models.WorkspaceKeyDetails + :type key: ~azure.synapse.artifacts.models.WorkspaceKeyDetails """ _validation = { @@ -9947,13 +10232,11 @@ class CustomEventsTrigger(MultiplePipelineTrigger): :type description: str :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRuntimeState + :vartype runtime_state: str or ~azure.synapse.artifacts.models.TriggerRuntimeState :param annotations: List of tags that can be used for describing the trigger. :type annotations: list[any] :param pipelines: Pipelines that need to be started. - :type pipelines: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerPipelineReference] + :type pipelines: list[~azure.synapse.artifacts.models.TriggerPipelineReference] :param subject_begins_with: The event subject must begin with the pattern provided for trigger to fire. At least one of these must be provided: subjectBeginsWith, subjectEndsWith. :type subject_begins_with: str @@ -10045,14 +10328,13 @@ class DatabricksNotebookActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param notebook_path: Required. The absolute path of the notebook to be run in the Databricks Workspace. This path must begin with a slash. Type: string (or Expression with resultType string). @@ -10110,14 +10392,13 @@ class DatabricksSparkJarActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param main_class_name: Required. The full name of the class containing the main method to be executed. This class must be contained in a JAR provided as a library. Type: string (or Expression with resultType string). @@ -10174,14 +10455,13 @@ class DatabricksSparkPythonActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param python_file: Required. The URI of the Python file to be executed. DBFS paths are supported. Type: string (or Expression with resultType string). :type python_file: any @@ -10238,7 +10518,7 @@ class DataFlow(msrest.serialization.Model): :type annotations: list[any] :param folder: The folder that this data flow is in. If not specified, Data flow will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowFolder + :type folder: ~azure.synapse.artifacts.models.DataFlowFolder """ _validation = { @@ -10267,31 +10547,59 @@ def __init__( self.folder = kwargs.get('folder', None) -class DataFlowDebugCommandRequest(msrest.serialization.Model): - """Request body structure for data flow expression preview. +class DataFlowDebugCommandPayload(msrest.serialization.Model): + """Structure of command payload. All required parameters must be populated in order to send to Azure. - :param session_id: Required. The ID of data flow debug session. - :type session_id: str - :param data_flow_name: The data flow which contains the debug session. - :type data_flow_name: str - :param command_name: The command name. - :type command_name: str - :param command_payload: Required. The command payload object. - :type command_payload: any + :param stream_name: Required. The stream name which is used for preview. + :type stream_name: str + :param row_limits: Row limits for preview response. + :type row_limits: int + :param columns: Array of column names. + :type columns: list[str] + :param expression: The expression which is used for preview. + :type expression: str """ _validation = { - 'session_id': {'required': True}, - 'command_payload': {'required': True}, + 'stream_name': {'required': True}, } + _attribute_map = { + 'stream_name': {'key': 'streamName', 'type': 'str'}, + 'row_limits': {'key': 'rowLimits', 'type': 'int'}, + 'columns': {'key': 'columns', 'type': '[str]'}, + 'expression': {'key': 'expression', 'type': 'str'}, + } + + def __init__( + self, + **kwargs + ): + super(DataFlowDebugCommandPayload, self).__init__(**kwargs) + self.stream_name = kwargs['stream_name'] + self.row_limits = kwargs.get('row_limits', None) + self.columns = kwargs.get('columns', None) + self.expression = kwargs.get('expression', None) + + +class DataFlowDebugCommandRequest(msrest.serialization.Model): + """Request body structure for data flow expression preview. + + :param session_id: The ID of data flow debug session. + :type session_id: str + :param command: The command type. Possible values include: "executePreviewQuery", + "executeStatisticsQuery", "executeExpressionQuery". + :type command: str or ~azure.synapse.artifacts.models.DataFlowDebugCommandType + :param command_payload: The command payload object. + :type command_payload: ~azure.synapse.artifacts.models.DataFlowDebugCommandPayload + """ + _attribute_map = { 'session_id': {'key': 'sessionId', 'type': 'str'}, - 'data_flow_name': {'key': 'dataFlowName', 'type': 'str'}, - 'command_name': {'key': 'commandName', 'type': 'str'}, - 'command_payload': {'key': 'commandPayload', 'type': 'object'}, + 'command': {'key': 'command', 'type': 'str'}, + 'command_payload': {'key': 'commandPayload', 'type': 'DataFlowDebugCommandPayload'}, } def __init__( @@ -10299,10 +10607,9 @@ def __init__( **kwargs ): super(DataFlowDebugCommandRequest, self).__init__(**kwargs) - self.session_id = kwargs['session_id'] - self.data_flow_name = kwargs.get('data_flow_name', None) - self.command_name = kwargs.get('command_name', None) - self.command_payload = kwargs['command_payload'] + self.session_id = kwargs.get('session_id', None) + self.command = kwargs.get('command', None) + self.command_payload = kwargs.get('command_payload', None) class DataFlowDebugCommandResponse(msrest.serialization.Model): @@ -10337,17 +10644,15 @@ class DataFlowDebugPackage(msrest.serialization.Model): :param session_id: The ID of data flow debug session. :type session_id: str :param data_flow: Data flow instance. - :type data_flow: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowDebugResource + :type data_flow: ~azure.synapse.artifacts.models.DataFlowDebugResource :param datasets: List of datasets. - :type datasets: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetDebugResource] + :type datasets: list[~azure.synapse.artifacts.models.DatasetDebugResource] :param linked_services: List of linked services. - :type linked_services: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceDebugResource] + :type linked_services: list[~azure.synapse.artifacts.models.LinkedServiceDebugResource] :param staging: Staging info for debug session. - :type staging: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowStagingInfo + :type staging: ~azure.synapse.artifacts.models.DataFlowStagingInfo :param debug_settings: Data flow debug settings. - :type debug_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowDebugPackageDebugSettings + :type debug_settings: ~azure.synapse.artifacts.models.DataFlowDebugPackageDebugSettings """ _attribute_map = { @@ -10378,8 +10683,7 @@ class DataFlowDebugPackageDebugSettings(msrest.serialization.Model): """Data flow debug settings. :param source_settings: Source setting for data flow debug. - :type source_settings: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowSourceSetting] + :type source_settings: list[~azure.synapse.artifacts.models.DataFlowSourceSetting] :param parameters: Data flow parameters. :type parameters: dict[str, any] :param dataset_parameters: Parameters for dataset. @@ -10479,7 +10783,7 @@ class DataFlowDebugResource(SubResourceDebugResource): :param name: The resource name. :type name: str :param properties: Required. Data flow properties. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlow + :type properties: ~azure.synapse.artifacts.models.DataFlow """ _validation = { @@ -10634,7 +10938,7 @@ class DataFlowListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of data flows. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowResource] + :type value: list[~azure.synapse.artifacts.models.DataFlowResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -10666,7 +10970,7 @@ class DataFlowReference(msrest.serialization.Model): collection. :type additional_properties: dict[str, any] :param type: Required. Data flow reference type. Possible values include: "DataFlowReference". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowReferenceType + :type type: str or ~azure.synapse.artifacts.models.DataFlowReferenceType :param reference_name: Required. Reference data flow name. :type reference_name: str :param dataset_parameters: Reference data flow parameters from dataset. @@ -10752,7 +11056,7 @@ class DataFlowResource(SubResource): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Data flow properties. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlow + :type properties: ~azure.synapse.artifacts.models.DataFlow """ _validation = { @@ -10818,13 +11122,11 @@ class DataFlowSink(Transformation): :param description: Transformation description. :type description: str :param dataset: Dataset reference. - :type dataset: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference + :type dataset: ~azure.synapse.artifacts.models.DatasetReference :param linked_service: Linked service reference. - :type linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param schema_linked_service: Schema linked service reference. - :type schema_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type schema_linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference """ _validation = { @@ -10859,13 +11161,11 @@ class DataFlowSource(Transformation): :param description: Transformation description. :type description: str :param dataset: Dataset reference. - :type dataset: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference + :type dataset: ~azure.synapse.artifacts.models.DatasetReference :param linked_service: Linked service reference. - :type linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param schema_linked_service: Schema linked service reference. - :type schema_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type schema_linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference """ _validation = { @@ -10922,8 +11222,7 @@ class DataFlowStagingInfo(msrest.serialization.Model): """Staging info for execute data flow activity. :param linked_service: Staging linked service reference. - :type linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param folder_path: Folder path for staging blob. :type folder_path: str """ @@ -10957,20 +11256,18 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param script_path: Required. Case-sensitive path to folder that contains the U-SQL script. Type: string (or Expression with resultType string). :type script_path: any :param script_linked_service: Required. Script linked service reference. - :type script_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type script_linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param degree_of_parallelism: The maximum number of nodes simultaneously used to run the job. Default value is 1. Type: integer (or Expression with resultType integer), minimum: 1. :type degree_of_parallelism: any @@ -11062,9 +11359,8 @@ class DatasetCompression(msrest.serialization.Model): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any + :param type: Required. Type of dataset compression.Constant filled by server. + :type type: str """ _validation = { @@ -11073,7 +11369,7 @@ class DatasetCompression(msrest.serialization.Model): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, } _subtype_map = { @@ -11097,9 +11393,8 @@ class DatasetBZip2Compression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any + :param type: Required. Type of dataset compression.Constant filled by server. + :type type: str """ _validation = { @@ -11108,7 +11403,7 @@ class DatasetBZip2Compression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, } def __init__( @@ -11150,7 +11445,7 @@ class DatasetDebugResource(SubResourceDebugResource): :param name: The resource name. :type name: str :param properties: Required. Dataset properties. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.Dataset + :type properties: ~azure.synapse.artifacts.models.Dataset """ _validation = { @@ -11178,10 +11473,10 @@ class DatasetDeflateCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - :param level: The Deflate compression level. + :param type: Required. Type of dataset compression.Constant filled by server. + :type type: str + :param level: The Deflate compression level. Type: string (or Expression with resultType + string). :type level: any """ @@ -11191,7 +11486,7 @@ class DatasetDeflateCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, 'level': {'key': 'level', 'type': 'object'}, } @@ -11231,10 +11526,9 @@ class DatasetGZipCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - :param level: The GZip compression level. + :param type: Required. Type of dataset compression.Constant filled by server. + :type type: str + :param level: The GZip compression level. Type: string (or Expression with resultType string). :type level: any """ @@ -11244,7 +11538,7 @@ class DatasetGZipCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, 'level': {'key': 'level', 'type': 'object'}, } @@ -11263,7 +11557,7 @@ class DatasetListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of datasets. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetResource] + :type value: list[~azure.synapse.artifacts.models.DatasetResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -11292,7 +11586,7 @@ class DatasetReference(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param type: Required. Dataset reference type. Possible values include: "DatasetReference". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReferenceType + :type type: str or ~azure.synapse.artifacts.models.DatasetReferenceType :param reference_name: Required. Reference dataset name. :type reference_name: str :param parameters: Arguments for dataset. @@ -11338,7 +11632,7 @@ class DatasetResource(SubResource): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Dataset properties. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.Dataset + :type properties: ~azure.synapse.artifacts.models.Dataset """ _validation = { @@ -11401,9 +11695,8 @@ class DatasetTarCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any + :param type: Required. Type of dataset compression.Constant filled by server. + :type type: str """ _validation = { @@ -11412,7 +11705,7 @@ class DatasetTarCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, } def __init__( @@ -11431,10 +11724,10 @@ class DatasetTarGZipCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - :param level: The TarGZip compression level. + :param type: Required. Type of dataset compression.Constant filled by server. + :type type: str + :param level: The TarGZip compression level. Type: string (or Expression with resultType + string). :type level: any """ @@ -11444,7 +11737,7 @@ class DatasetTarGZipCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, 'level': {'key': 'level', 'type': 'object'}, } @@ -11465,10 +11758,10 @@ class DatasetZipDeflateCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - :param level: The ZipDeflate compression level. + :param type: Required. Type of dataset compression.Constant filled by server. + :type type: str + :param level: The ZipDeflate compression level. Type: string (or Expression with resultType + string). :type level: any """ @@ -11478,7 +11771,7 @@ class DatasetZipDeflateCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, 'level': {'key': 'level', 'type': 'object'}, } @@ -11502,13 +11795,11 @@ class Db2LinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: The connection string. It is mutually exclusive with server, @@ -11523,13 +11814,12 @@ class Db2LinkedService(LinkedService): :type database: any :param authentication_type: AuthenticationType to be used for connection. It is mutually exclusive with connectionString property. Possible values include: "Basic". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.Db2AuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.Db2AuthenticationType :param username: Username for authentication. It is mutually exclusive with connectionString property. Type: string (or Expression with resultType string). :type username: any :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param package_collection: Under where packages are created when querying database. It is mutually exclusive with connectionString property. Type: string (or Expression with resultType string). @@ -11656,16 +11946,14 @@ class Db2TableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -11722,14 +12010,13 @@ class DeleteActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param recursive: If true, files or sub-folders under current folder path will be deleted recursively. Default is false. Type: boolean (or Expression with resultType boolean). :type recursive: any @@ -11741,12 +12028,11 @@ class DeleteActivity(ExecutionActivity): :type enable_logging: any :param log_storage_settings: Log storage settings customer need to provide when enableLogging is true. - :type log_storage_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LogStorageSettings + :type log_storage_settings: ~azure.synapse.artifacts.models.LogStorageSettings :param dataset: Required. Delete activity dataset reference. - :type dataset: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference + :type dataset: ~azure.synapse.artifacts.models.DatasetReference :param store_settings: Delete activity store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings """ _validation = { @@ -11829,18 +12115,16 @@ class DelimitedTextDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param location: The location of the delimited text storage. - :type location: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetLocation + :type location: ~azure.synapse.artifacts.models.DatasetLocation :param column_delimiter: The column delimiter. Type: string (or Expression with resultType string). :type column_delimiter: any @@ -11852,10 +12136,9 @@ class DelimitedTextDataset(Dataset): https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string (or Expression with resultType string). :type encoding_name: any - :param compression_codec: Possible values include: "bzip2", "gzip", "deflate", "zipDeflate", - "snappy", "lz4", "tar", "tarGZip". - :type compression_codec: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.CompressionCodec + :param compression_codec: The data compressionCodec. Type: string (or Expression with + resultType string). + :type compression_codec: any :param compression_level: The data compression method used for DelimitedText. :type compression_level: any :param quote_char: The quote character. Type: string (or Expression with resultType string). @@ -11889,7 +12172,7 @@ class DelimitedTextDataset(Dataset): 'column_delimiter': {'key': 'typeProperties.columnDelimiter', 'type': 'object'}, 'row_delimiter': {'key': 'typeProperties.rowDelimiter', 'type': 'object'}, 'encoding_name': {'key': 'typeProperties.encodingName', 'type': 'object'}, - 'compression_codec': {'key': 'typeProperties.compressionCodec', 'type': 'str'}, + 'compression_codec': {'key': 'typeProperties.compressionCodec', 'type': 'object'}, 'compression_level': {'key': 'typeProperties.compressionLevel', 'type': 'object'}, 'quote_char': {'key': 'typeProperties.quoteChar', 'type': 'object'}, 'escape_char': {'key': 'typeProperties.escapeChar', 'type': 'object'}, @@ -11929,8 +12212,7 @@ class DelimitedTextReadSettings(FormatReadSettings): input files. Type: integer (or Expression with resultType integer). :type skip_line_count: any :param compression_properties: Compression settings. - :type compression_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.CompressionReadSettings + :type compression_properties: ~azure.synapse.artifacts.models.CompressionReadSettings """ _validation = { @@ -11980,10 +12262,9 @@ class DelimitedTextSink(CopySink): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: DelimitedText store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreWriteSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreWriteSettings :param format_settings: DelimitedText format settings. - :type format_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.DelimitedTextWriteSettings + :type format_settings: ~azure.synapse.artifacts.models.DelimitedTextWriteSettings """ _validation = { @@ -12032,10 +12313,9 @@ class DelimitedTextSource(CopySource): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: DelimitedText store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings :param format_settings: DelimitedText format settings. - :type format_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.DelimitedTextReadSettings + :type format_settings: ~azure.synapse.artifacts.models.DelimitedTextReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :type additional_columns: any @@ -12207,16 +12487,14 @@ class DocumentDbCollectionDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param collection_name: Required. Document Database collection name. Type: string (or Expression with resultType string). :type collection_name: any @@ -12380,20 +12658,18 @@ class DrillLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type pwd: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -12500,16 +12776,14 @@ class DrillTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -12582,8 +12856,7 @@ class DWCopyCommandSettings(msrest.serialization.Model): default values in the property overwrite the DEFAULT constraint set in the DB, and identity column cannot have a default value. Type: array of objects (or Expression with resultType array of objects). - :type default_values: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.DWCopyCommandDefaultValue] + :type default_values: list[~azure.synapse.artifacts.models.DWCopyCommandDefaultValue] :param additional_options: Additional options directly passed to SQL DW in Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalOptions": { "MAXERRORS": "1000", "DATEFORMAT": "'ymd'" }. @@ -12634,13 +12907,11 @@ class DynamicsAXLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param url: Required. The Dynamics AX (or Dynamics 365 Finance and Operations) instance OData @@ -12652,7 +12923,7 @@ class DynamicsAXLinkedService(LinkedService): :param service_principal_key: Required. Specify the application's key. Mark this field as a SecureString to store it securely in Data Factory, or reference a secret stored in Azure Key Vault. Type: string (or Expression with resultType string). - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: Required. Specify the tenant information (domain name or tenant ID) under which your application resides. Retrieve it by hovering the mouse in the top-right corner of the Azure portal. Type: string (or Expression with resultType string). @@ -12723,16 +12994,14 @@ class DynamicsAXResourceDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param path: Required. The path of the Dynamics AX OData entity. Type: string (or Expression with resultType string). :type path: any @@ -12846,16 +13115,14 @@ class DynamicsCrmEntityDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param entity_name: The logical name of the entity. Type: string (or Expression with resultType string). :type entity_name: any @@ -12899,21 +13166,17 @@ class DynamicsCrmLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param deployment_type: Required. The deployment type of the Dynamics CRM instance. 'Online' for Dynamics CRM Online and 'OnPremisesWithIfd' for Dynamics CRM on-premises with Ifd. Type: - string (or Expression with resultType string). Possible values include: "Online", - "OnPremisesWithIfd". - :type deployment_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsDeploymentType + string (or Expression with resultType string). + :type deployment_type: any :param host_name: The host name of the on-premises Dynamics CRM server. The property is required for on-prem and not allowed for online. Type: string (or Expression with resultType string). @@ -12932,15 +13195,13 @@ class DynamicsCrmLinkedService(LinkedService): :param authentication_type: Required. The authentication type to connect to Dynamics CRM server. 'Office365' for online scenario, 'Ifd' for on-premises with Ifd scenario, 'AADServicePrincipal' for Server-To-Server authentication in online scenario. Type: string (or - Expression with resultType string). Possible values include: "Office365", "Ifd", - "AADServicePrincipal". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsAuthenticationType + Expression with resultType string). + :type authentication_type: any :param username: User name to access the Dynamics CRM instance. Type: string (or Expression with resultType string). :type username: any :param password: Password to access the Dynamics CRM instance. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param service_principal_id: The client ID of the application in Azure Active Directory used for Server-To-Server authentication. Type: string (or Expression with resultType string). :type service_principal_id: any @@ -12952,8 +13213,7 @@ class DynamicsCrmLinkedService(LinkedService): servicePrincipalCredential can be SecureString or AzureKeyVaultSecretReference. If servicePrincipalCredentialType is 'ServicePrincipalCert', servicePrincipalCredential can only be AzureKeyVaultSecretReference. - :type service_principal_credential: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_credential: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -12973,12 +13233,12 @@ class DynamicsCrmLinkedService(LinkedService): 'description': {'key': 'description', 'type': 'str'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'str'}, + 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'object'}, 'host_name': {'key': 'typeProperties.hostName', 'type': 'object'}, 'port': {'key': 'typeProperties.port', 'type': 'object'}, 'service_uri': {'key': 'typeProperties.serviceUri', 'type': 'object'}, 'organization_name': {'key': 'typeProperties.organizationName', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, 'username': {'key': 'typeProperties.username', 'type': 'object'}, 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, @@ -13034,8 +13294,7 @@ class DynamicsCrmSink(CopySink): :type max_concurrent_connections: any :param write_behavior: Required. The write behavior for the operation. Possible values include: "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsSinkWriteBehavior + :type write_behavior: str or ~azure.synapse.artifacts.models.DynamicsSinkWriteBehavior :param ignore_null_values: The flag indicating whether to ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). @@ -13144,16 +13403,14 @@ class DynamicsEntityDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param entity_name: The logical name of the entity. Type: string (or Expression with resultType string). :type entity_name: any @@ -13197,20 +13454,17 @@ class DynamicsLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param deployment_type: Required. The deployment type of the Dynamics instance. 'Online' for Dynamics Online and 'OnPremisesWithIfd' for Dynamics on-premises with Ifd. Type: string (or - Expression with resultType string). Possible values include: "Online", "OnPremisesWithIfd". - :type deployment_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsDeploymentType + Expression with resultType string). + :type deployment_type: any :param host_name: The host name of the on-premises Dynamics server. The property is required for on-prem and not allowed for online. Type: string (or Expression with resultType string). :type host_name: any @@ -13228,30 +13482,26 @@ class DynamicsLinkedService(LinkedService): :param authentication_type: Required. The authentication type to connect to Dynamics server. 'Office365' for online scenario, 'Ifd' for on-premises with Ifd scenario, 'AADServicePrincipal' for Server-To-Server authentication in online scenario. Type: string (or Expression with - resultType string). Possible values include: "Office365", "Ifd", "AADServicePrincipal". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsAuthenticationType + resultType string). + :type authentication_type: any :param username: User name to access the Dynamics instance. Type: string (or Expression with resultType string). :type username: any :param password: Password to access the Dynamics instance. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param service_principal_id: The client ID of the application in Azure Active Directory used for Server-To-Server authentication. Type: string (or Expression with resultType string). :type service_principal_id: any :param service_principal_credential_type: The service principal credential type to use in Server-To-Server authentication. 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' - for certificate. Type: string (or Expression with resultType string). Possible values include: - "ServicePrincipalKey", "ServicePrincipalCert". - :type service_principal_credential_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsServicePrincipalCredentialType + for certificate. Type: string (or Expression with resultType string). + :type service_principal_credential_type: any :param service_principal_credential: The credential of the service principal object in Azure Active Directory. If servicePrincipalCredentialType is 'ServicePrincipalKey', servicePrincipalCredential can be SecureString or AzureKeyVaultSecretReference. If servicePrincipalCredentialType is 'ServicePrincipalCert', servicePrincipalCredential can only be AzureKeyVaultSecretReference. - :type service_principal_credential: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_credential: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -13271,16 +13521,16 @@ class DynamicsLinkedService(LinkedService): 'description': {'key': 'description', 'type': 'str'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'str'}, + 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'object'}, 'host_name': {'key': 'typeProperties.hostName', 'type': 'object'}, 'port': {'key': 'typeProperties.port', 'type': 'object'}, 'service_uri': {'key': 'typeProperties.serviceUri', 'type': 'object'}, 'organization_name': {'key': 'typeProperties.organizationName', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, 'username': {'key': 'typeProperties.username', 'type': 'object'}, 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_credential_type': {'key': 'typeProperties.servicePrincipalCredentialType', 'type': 'str'}, + 'service_principal_credential_type': {'key': 'typeProperties.servicePrincipalCredentialType', 'type': 'object'}, 'service_principal_credential': {'key': 'typeProperties.servicePrincipalCredential', 'type': 'SecretBase'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } @@ -13332,8 +13582,7 @@ class DynamicsSink(CopySink): :type max_concurrent_connections: any :param write_behavior: Required. The write behavior for the operation. Possible values include: "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsSinkWriteBehavior + :type write_behavior: str or ~azure.synapse.artifacts.models.DynamicsSinkWriteBehavior :param ignore_null_values: The flag indicating whether ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). @@ -13434,13 +13683,11 @@ class EloquaLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param endpoint: Required. The endpoint of the Eloqua server. (i.e. eloqua.example.com). @@ -13449,7 +13696,7 @@ class EloquaLinkedService(LinkedService): sitename/username. (i.e. Eloqua/Alice). :type username: any :param password: The password corresponding to the user name. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -13522,16 +13769,14 @@ class EloquaObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -13625,7 +13870,7 @@ class EncryptionDetails(msrest.serialization.Model): :ivar double_encryption_enabled: Double Encryption enabled. :vartype double_encryption_enabled: bool :param cmk: Customer Managed Key Details. - :type cmk: ~azure.synapse.artifacts.v2021_06_01_preview.models.CustomerManagedKeyDetails + :type cmk: ~azure.synapse.artifacts.models.CustomerManagedKeyDetails """ _validation = { @@ -13651,8 +13896,7 @@ class EntityReference(msrest.serialization.Model): :param type: The type of this referenced entity. Possible values include: "IntegrationRuntimeReference", "LinkedServiceReference". - :type type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeEntityReferenceType + :type type: str or ~azure.synapse.artifacts.models.IntegrationRuntimeEntityReferenceType :param reference_name: The name of this referenced entity. :type reference_name: str """ @@ -13705,7 +13949,7 @@ class ErrorContract(msrest.serialization.Model): """Contains details when the response code indicates an error. :param error: The error details. - :type error: ~azure.synapse.artifacts.v2021_06_01_preview.models.ErrorResponse + :type error: ~azure.synapse.artifacts.models.ErrorResponse """ _attribute_map = { @@ -13732,10 +13976,9 @@ class ErrorResponse(msrest.serialization.Model): :ivar target: The error target. :vartype target: str :ivar details: The error details. - :vartype details: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ErrorResponse] + :vartype details: list[~azure.synapse.artifacts.models.ErrorResponse] :ivar additional_info: The error additional info. - :vartype additional_info: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.ErrorAdditionalInfo] + :vartype additional_info: list[~azure.synapse.artifacts.models.ErrorAdditionalInfo] """ _validation = { @@ -13820,18 +14063,16 @@ class ExcelDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param location: The location of the excel storage. - :type location: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetLocation + :type location: ~azure.synapse.artifacts.models.DatasetLocation :param sheet_name: The sheet of excel file. Type: string (or Expression with resultType string). :type sheet_name: any @@ -13843,7 +14084,7 @@ class ExcelDataset(Dataset): false. Type: boolean (or Expression with resultType boolean). :type first_row_as_header: any :param compression: The data compression method used for the json dataset. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression :param null_value: The null value string. Type: string (or Expression with resultType string). :type null_value: any """ @@ -13905,7 +14146,7 @@ class ExcelSource(CopySource): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Excel store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :type additional_columns: any @@ -13950,24 +14191,21 @@ class ExecuteDataFlowActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param dataflow: Required. Data flow reference. - :type dataflow: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowReference + :type dataflow: ~azure.synapse.artifacts.models.DataFlowReference :param staging: Staging info for execute data flow activity. - :type staging: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowStagingInfo + :type staging: ~azure.synapse.artifacts.models.DataFlowStagingInfo :param integration_runtime: The integration runtime reference. - :type integration_runtime: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type integration_runtime: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param compute: Compute properties for data flow activity. - :type compute: - ~azure.synapse.artifacts.v2021_06_01_preview.models.ExecuteDataFlowActivityTypePropertiesCompute + :type compute: ~azure.synapse.artifacts.models.ExecuteDataFlowActivityTypePropertiesCompute :param trace_level: Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string). :type trace_level: any @@ -14024,8 +14262,7 @@ class ExecuteDataFlowActivityTypePropertiesCompute(msrest.serialization.Model): :param compute_type: Compute type of the cluster which will execute data flow job. Possible values include: "General", "MemoryOptimized", "ComputeOptimized". - :type compute_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowComputeType + :type compute_type: str or ~azure.synapse.artifacts.models.DataFlowComputeType :param core_count: Core count of the cluster which will execute data flow job. Supported values are: 8, 16, 32, 48, 80, 144 and 272. :type core_count: int @@ -14060,11 +14297,11 @@ class ExecutePipelineActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param pipeline: Required. Pipeline reference. - :type pipeline: ~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineReference + :type pipeline: ~azure.synapse.artifacts.models.PipelineReference :param parameters: Pipeline parameters. :type parameters: dict[str, any] :param wait_on_completion: Defines whether activity execution will wait for the dependent @@ -14116,16 +14353,15 @@ class ExecuteSSISPackageActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param package_location: Required. SSIS package location. - :type package_location: ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISPackageLocation + :type package_location: ~azure.synapse.artifacts.models.SSISPackageLocation :param runtime: Specifies the runtime to execute SSIS package. The value should be "x86" or "x64". Type: string (or Expression with resultType string). :type runtime: any @@ -14136,30 +14372,25 @@ class ExecuteSSISPackageActivity(ExecutionActivity): Expression with resultType string). :type environment_path: any :param execution_credential: The package execution credential. - :type execution_credential: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISExecutionCredential + :type execution_credential: ~azure.synapse.artifacts.models.SSISExecutionCredential :param connect_via: Required. The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param project_parameters: The project level parameters to execute the SSIS package. - :type project_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISExecutionParameter] + :type project_parameters: dict[str, ~azure.synapse.artifacts.models.SSISExecutionParameter] :param package_parameters: The package level parameters to execute the SSIS package. - :type package_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISExecutionParameter] + :type package_parameters: dict[str, ~azure.synapse.artifacts.models.SSISExecutionParameter] :param project_connection_managers: The project level connection managers to execute the SSIS package. :type project_connection_managers: dict[str, dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISExecutionParameter]] + ~azure.synapse.artifacts.models.SSISExecutionParameter]] :param package_connection_managers: The package level connection managers to execute the SSIS package. :type package_connection_managers: dict[str, dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISExecutionParameter]] + ~azure.synapse.artifacts.models.SSISExecutionParameter]] :param property_overrides: The property overrides to execute the SSIS package. - :type property_overrides: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISPropertyOverride] + :type property_overrides: dict[str, ~azure.synapse.artifacts.models.SSISPropertyOverride] :param log_location: SSIS package execution log location. - :type log_location: ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISLogLocation + :type log_location: ~azure.synapse.artifacts.models.SSISLogLocation """ _validation = { @@ -14271,7 +14502,7 @@ class Expression(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param type: Required. Expression type. Possible values include: "Expression". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.ExpressionType + :type type: str or ~azure.synapse.artifacts.models.ExpressionType :param value: Required. Expression value. :type value: str """ @@ -14306,13 +14537,11 @@ class FileServerLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. Host name of the server. Type: string (or Expression with resultType @@ -14322,7 +14551,7 @@ class FileServerLinkedService(LinkedService): string). :type user_id: any :param password: Password to logon the server. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -14534,16 +14763,14 @@ class FileShareDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param folder_path: The path of the on-premises file system. Type: string (or Expression with resultType string). :type folder_path: any @@ -14557,12 +14784,12 @@ class FileShareDataset(Dataset): with resultType string). :type modified_datetime_end: any :param format: The format of the files. - :type format: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetStorageFormat + :type format: ~azure.synapse.artifacts.models.DatasetStorageFormat :param file_filter: Specify a filter to be used to select a subset of files in the folderPath rather than all files. Type: string (or Expression with resultType string). :type file_filter: any :param compression: The data compression method used for the file system. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression """ _validation = { @@ -14723,13 +14950,13 @@ class FilterActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param items: Required. Input array on which filter should be applied. - :type items: ~azure.synapse.artifacts.v2021_06_01_preview.models.Expression + :type items: ~azure.synapse.artifacts.models.Expression :param condition: Required. Condition to be used for filtering the input. - :type condition: ~azure.synapse.artifacts.v2021_06_01_preview.models.Expression + :type condition: ~azure.synapse.artifacts.models.Expression """ _validation = { @@ -14775,18 +15002,18 @@ class ForEachActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param is_sequential: Should the loop be executed in sequence or in parallel (max 50). :type is_sequential: bool :param batch_count: Batch count to be used for controlling the number of parallel execution (when isSequential is set to false). :type batch_count: int :param items: Required. Collection to iterate. - :type items: ~azure.synapse.artifacts.v2021_06_01_preview.models.Expression + :type items: ~azure.synapse.artifacts.models.Expression :param activities: Required. List of activities to execute . - :type activities: list[~azure.synapse.artifacts.v2021_06_01_preview.models.Activity] + :type activities: list[~azure.synapse.artifacts.models.Activity] """ _validation = { @@ -14905,13 +15132,11 @@ class FtpServerLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. Host name of the FTP server. Type: string (or Expression with resultType @@ -14922,13 +15147,12 @@ class FtpServerLinkedService(LinkedService): :type port: any :param authentication_type: The authentication type to be used to connect to the FTP server. Possible values include: "Basic", "Anonymous". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.FtpAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.FtpAuthenticationType :param user_name: Username to logon the FTP server. Type: string (or Expression with resultType string). :type user_name: any :param password: Password to logon the FTP server. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -15032,22 +15256,21 @@ class GetMetadataActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param dataset: Required. GetMetadata activity dataset reference. - :type dataset: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference + :type dataset: ~azure.synapse.artifacts.models.DatasetReference :param field_list: Fields of metadata to get from dataset. :type field_list: list[any] :param store_settings: GetMetadata activity store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings :param format_settings: GetMetadata activity format settings. - :type format_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.FormatReadSettings + :type format_settings: ~azure.synapse.artifacts.models.FormatReadSettings """ _validation = { @@ -15190,13 +15413,11 @@ class GoogleAdWordsLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param client_customer_id: Required. The Client customer ID of the AdWords account that you @@ -15204,21 +15425,21 @@ class GoogleAdWordsLinkedService(LinkedService): :type client_customer_id: any :param developer_token: Required. The developer token associated with the manager account that you use to grant access to the AdWords API. - :type developer_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type developer_token: ~azure.synapse.artifacts.models.SecretBase :param authentication_type: Required. The OAuth 2.0 authentication mechanism used for authentication. ServiceAuthentication can only be used on self-hosted IR. Possible values include: "ServiceAuthentication", "UserAuthentication". :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.GoogleAdWordsAuthenticationType + ~azure.synapse.artifacts.models.GoogleAdWordsAuthenticationType :param refresh_token: The refresh token obtained from Google for authorizing access to AdWords for UserAuthentication. - :type refresh_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type refresh_token: ~azure.synapse.artifacts.models.SecretBase :param client_id: The client id of the google application used to acquire the refresh token. Type: string (or Expression with resultType string). :type client_id: any :param client_secret: The client secret of the google application used to acquire the refresh token. - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type client_secret: ~azure.synapse.artifacts.models.SecretBase :param email: The service account email ID that is used for ServiceAuthentication and can only be used on self-hosted IR. :type email: any @@ -15303,16 +15524,14 @@ class GoogleAdWordsObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -15409,13 +15628,11 @@ class GoogleBigQueryLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param project: Required. The default BigQuery project to query against. @@ -15430,16 +15647,16 @@ class GoogleBigQueryLinkedService(LinkedService): authentication. ServiceAuthentication can only be used on self-hosted IR. Possible values include: "ServiceAuthentication", "UserAuthentication". :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.GoogleBigQueryAuthenticationType + ~azure.synapse.artifacts.models.GoogleBigQueryAuthenticationType :param refresh_token: The refresh token obtained from Google for authorizing access to BigQuery for UserAuthentication. - :type refresh_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type refresh_token: ~azure.synapse.artifacts.models.SecretBase :param client_id: The client id of the google application used to acquire the refresh token. Type: string (or Expression with resultType string). :type client_id: any :param client_secret: The client secret of the google application used to acquire the refresh token. - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type client_secret: ~azure.synapse.artifacts.models.SecretBase :param email: The service account email ID that is used for ServiceAuthentication and can only be used on self-hosted IR. :type email: any @@ -15525,16 +15742,14 @@ class GoogleBigQueryObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using database + table properties instead. :type table_name: any @@ -15642,13 +15857,11 @@ class GoogleCloudStorageLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param access_key_id: The access key identifier of the Google Cloud Storage Identity and Access @@ -15656,7 +15869,7 @@ class GoogleCloudStorageLinkedService(LinkedService): :type access_key_id: any :param secret_access_key: The secret access key of the Google Cloud Storage Identity and Access Management (IAM) user. - :type secret_access_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type secret_access_key: ~azure.synapse.artifacts.models.SecretBase :param service_url: This value specifies the endpoint to access with the Google Cloud Storage Connector. This is an optional property; change it only if you want to try a different service endpoint or want to switch between https and http. Type: string (or Expression with resultType @@ -15838,20 +16051,18 @@ class GreenplumLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type pwd: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -15958,16 +16169,14 @@ class GreenplumTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -16020,13 +16229,11 @@ class HBaseLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The IP address or host name of the HBase server. (i.e. 192.168.222.160). @@ -16039,12 +16246,11 @@ class HBaseLinkedService(LinkedService): :type http_path: any :param authentication_type: Required. The authentication mechanism to use to connect to the HBase server. Possible values include: "Anonymous", "Basic". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.HBaseAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.HBaseAuthenticationType :param username: The user name used to connect to the HBase instance. :type username: any :param password: The password corresponding to the user name. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The default value is false. :type enable_ssl: any @@ -16128,16 +16334,14 @@ class HBaseObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -16234,13 +16438,11 @@ class HdfsLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param url: Required. The URL of the HDFS service endpoint, e.g. @@ -16257,7 +16459,7 @@ class HdfsLinkedService(LinkedService): resultType string). :type user_name: any :param password: Password for Windows authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase """ _validation = { @@ -16367,7 +16569,7 @@ class HdfsReadSettings(StoreReadSettings): with resultType string). :type modified_datetime_end: any :param distcp_settings: Specifies Distcp-related settings. - :type distcp_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.DistcpSettings + :type distcp_settings: ~azure.synapse.artifacts.models.DistcpSettings :param delete_files_after_completion: Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). :type delete_files_after_completion: any @@ -16434,7 +16636,7 @@ class HdfsSource(CopySource): true. Type: boolean (or Expression with resultType boolean). :type recursive: any :param distcp_settings: Specifies Distcp-related settings. - :type distcp_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.DistcpSettings + :type distcp_settings: ~azure.synapse.artifacts.models.DistcpSettings """ _validation = { @@ -16476,27 +16678,23 @@ class HDInsightHiveActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param storage_linked_services: Storage linked service references. - :type storage_linked_services: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference] + :type storage_linked_services: list[~azure.synapse.artifacts.models.LinkedServiceReference] :param arguments: User specified arguments to HDInsightActivity. :type arguments: list[any] :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.HDInsightActivityDebugInfoOption + :type get_debug_info: str or ~azure.synapse.artifacts.models.HDInsightActivityDebugInfoOption :param script_path: Script path. Type: string (or Expression with resultType string). :type script_path: any :param script_linked_service: Script linked service reference. - :type script_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type script_linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param defines: Allows user to specify defines for Hive job request. :type defines: dict[str, any] :param variables: User specified arguments under hivevar namespace. @@ -16557,13 +16755,11 @@ class HDInsightLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param cluster_uri: Required. HDInsight cluster URI. Type: string (or Expression with @@ -16573,14 +16769,12 @@ class HDInsightLinkedService(LinkedService): string). :type user_name: any :param password: HDInsight cluster password. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param linked_service_name: The Azure Storage linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param hcatalog_linked_service_name: A reference to the Azure SQL linked service that points to the HCatalog database. - :type hcatalog_linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type hcatalog_linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -16646,29 +16840,25 @@ class HDInsightMapReduceActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param storage_linked_services: Storage linked service references. - :type storage_linked_services: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference] + :type storage_linked_services: list[~azure.synapse.artifacts.models.LinkedServiceReference] :param arguments: User specified arguments to HDInsightActivity. :type arguments: list[any] :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.HDInsightActivityDebugInfoOption + :type get_debug_info: str or ~azure.synapse.artifacts.models.HDInsightActivityDebugInfoOption :param class_name: Required. Class name. Type: string (or Expression with resultType string). :type class_name: any :param jar_file_path: Required. Jar path. Type: string (or Expression with resultType string). :type jar_file_path: any :param jar_linked_service: Jar linked service reference. - :type jar_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type jar_linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param jar_libs: Jar libs. :type jar_libs: list[any] :param defines: Allows user to specify defines for the MapReduce job request. @@ -16728,13 +16918,11 @@ class HDInsightOnDemandLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param cluster_size: Required. Number of worker/data nodes in the cluster. Suggestion value: 4. @@ -16750,8 +16938,7 @@ class HDInsightOnDemandLinkedService(LinkedService): :type version: any :param linked_service_name: Required. Azure Storage linked service to be used by the on-demand cluster for storing and processing data. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param host_subscription_id: Required. The customer’s subscription to host the cluster. Type: string (or Expression with resultType string). :type host_subscription_id: any @@ -16759,7 +16946,7 @@ class HDInsightOnDemandLinkedService(LinkedService): (or Expression with resultType string). :type service_principal_id: any :param service_principal_key: The key for the service principal id. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: Required. The Tenant id/name to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -16773,21 +16960,20 @@ class HDInsightOnDemandLinkedService(LinkedService): resultType string). :type cluster_user_name: any :param cluster_password: The password to access the cluster. - :type cluster_password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type cluster_password: ~azure.synapse.artifacts.models.SecretBase :param cluster_ssh_user_name: The username to SSH remotely connect to cluster’s node (for Linux). Type: string (or Expression with resultType string). :type cluster_ssh_user_name: any :param cluster_ssh_password: The password to SSH remotely connect cluster’s node (for Linux). - :type cluster_ssh_password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type cluster_ssh_password: ~azure.synapse.artifacts.models.SecretBase :param additional_linked_service_names: Specifies additional storage accounts for the HDInsight linked service so that the Data Factory service can register them on your behalf. :type additional_linked_service_names: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference] + list[~azure.synapse.artifacts.models.LinkedServiceReference] :param hcatalog_linked_service_name: The name of Azure SQL linked service that point to the HCatalog database. The on-demand HDInsight cluster is created by using the Azure SQL database as the metastore. - :type hcatalog_linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type hcatalog_linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param cluster_type: The cluster type. Type: string (or Expression with resultType string). :type cluster_type: any :param spark_version: The version of spark if the cluster type is 'spark'. Type: string (or @@ -16831,7 +17017,7 @@ class HDInsightOnDemandLinkedService(LinkedService): :param script_actions: Custom script actions to run on HDI ondemand cluster once it's up. Please refer to https://docs.microsoft.com/en-us/azure/hdinsight/hdinsight-hadoop-customize-cluster-linux?toc=%2Fen-us%2Fazure%2Fhdinsight%2Fr-server%2FTOC.json&bc=%2Fen-us%2Fazure%2Fbread%2Ftoc.json#understanding-script-actions. - :type script_actions: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ScriptAction] + :type script_actions: list[~azure.synapse.artifacts.models.ScriptAction] :param virtual_network_id: The ARM resource ID for the vNet to which the cluster should be joined after creation. Type: string (or Expression with resultType string). :type virtual_network_id: any @@ -16949,28 +17135,24 @@ class HDInsightPigActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param storage_linked_services: Storage linked service references. - :type storage_linked_services: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference] + :type storage_linked_services: list[~azure.synapse.artifacts.models.LinkedServiceReference] :param arguments: User specified arguments to HDInsightActivity. Type: array (or Expression with resultType array). :type arguments: any :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.HDInsightActivityDebugInfoOption + :type get_debug_info: str or ~azure.synapse.artifacts.models.HDInsightActivityDebugInfoOption :param script_path: Script path. Type: string (or Expression with resultType string). :type script_path: any :param script_linked_service: Script linked service reference. - :type script_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type script_linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param defines: Allows user to specify defines for Pig job request. :type defines: dict[str, any] """ @@ -17026,14 +17208,13 @@ class HDInsightSparkActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param root_path: Required. The root path in 'sparkJobLinkedService' for all the job’s files. Type: string (or Expression with resultType string). :type root_path: any @@ -17043,12 +17224,10 @@ class HDInsightSparkActivity(ExecutionActivity): :param arguments: The user-specified arguments to HDInsightSparkActivity. :type arguments: list[any] :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.HDInsightActivityDebugInfoOption + :type get_debug_info: str or ~azure.synapse.artifacts.models.HDInsightActivityDebugInfoOption :param spark_job_linked_service: The storage linked service for uploading the entry file and dependencies, and for receiving logs. - :type spark_job_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type spark_job_linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param class_name: The application's Java/Spark main class. :type class_name: str :param proxy_user: The user to impersonate that will execute the job. Type: string (or @@ -17115,22 +17294,19 @@ class HDInsightStreamingActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param storage_linked_services: Storage linked service references. - :type storage_linked_services: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference] + :type storage_linked_services: list[~azure.synapse.artifacts.models.LinkedServiceReference] :param arguments: User specified arguments to HDInsightActivity. :type arguments: list[any] :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.HDInsightActivityDebugInfoOption + :type get_debug_info: str or ~azure.synapse.artifacts.models.HDInsightActivityDebugInfoOption :param mapper: Required. Mapper executable name. Type: string (or Expression with resultType string). :type mapper: any @@ -17144,8 +17320,7 @@ class HDInsightStreamingActivity(ExecutionActivity): :param file_paths: Required. Paths to streaming job files. Can be directories. :type file_paths: list[any] :param file_linked_service: Linked service reference where the files are located. - :type file_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type file_linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param combiner: Combiner executable name. Type: string (or Expression with resultType string). :type combiner: any :param command_environment: Command line environment values. @@ -17218,13 +17393,11 @@ class HiveLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. IP address or host name of the Hive server, separated by ';' for @@ -17234,16 +17407,15 @@ class HiveLinkedService(LinkedService): :type port: any :param server_type: The type of Hive server. Possible values include: "HiveServer1", "HiveServer2", "HiveThriftServer". - :type server_type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.HiveServerType + :type server_type: str or ~azure.synapse.artifacts.models.HiveServerType :param thrift_transport_protocol: The transport protocol to use in the Thrift layer. Possible values include: "Binary", "SASL", "HTTP ". :type thrift_transport_protocol: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.HiveThriftTransportProtocol + ~azure.synapse.artifacts.models.HiveThriftTransportProtocol :param authentication_type: Required. The authentication method used to access the Hive server. Possible values include: "Anonymous", "Username", "UsernameAndPassword", "WindowsAzureHDInsightService". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.HiveAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.HiveAuthenticationType :param service_discovery_mode: true to indicate using the ZooKeeper service, false not. :type service_discovery_mode: any :param zoo_keeper_name_space: The namespace on ZooKeeper under which Hive Server 2 nodes are @@ -17256,7 +17428,7 @@ class HiveLinkedService(LinkedService): :type username: any :param password: The password corresponding to the user name that you provided in the Username field. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param http_path: The partial URL corresponding to the Hive server. :type http_path: any :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The @@ -17357,16 +17529,14 @@ class HiveObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -17481,16 +17651,14 @@ class HttpDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param relative_url: The relative URL based on the URL in the HttpLinkedService refers to an HTTP file Type: string (or Expression with resultType string). :type relative_url: any @@ -17507,9 +17675,9 @@ class HttpDataset(Dataset): string). :type additional_headers: any :param format: The format of files. - :type format: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetStorageFormat + :type format: ~azure.synapse.artifacts.models.DatasetStorageFormat :param compression: The data compression method used on files. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression """ _validation = { @@ -17560,13 +17728,11 @@ class HttpLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param url: Required. The base URL of the HTTP endpoint, e.g. http://www.microsoft.com. Type: @@ -17574,14 +17740,13 @@ class HttpLinkedService(LinkedService): :type url: any :param authentication_type: The authentication type to be used to connect to the HTTP server. Possible values include: "Basic", "Anonymous", "Digest", "Windows", "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.HttpAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.HttpAuthenticationType :param user_name: User name for Basic, Digest, or Windows authentication. Type: string (or Expression with resultType string). :type user_name: any :param password: Password for Basic, Digest, Windows, or ClientCertificate with EmbeddedCertData authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param embedded_cert_data: Base64 encoded certificate data for ClientCertificate authentication. For on-premises copy with ClientCertificate authentication, either CertThumbprint or EmbeddedCertData/Password should be specified. Type: string (or Expression @@ -17802,25 +17967,23 @@ class HubspotLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param client_id: Required. The client ID associated with your Hubspot application. :type client_id: any :param client_secret: The client secret associated with your Hubspot application. - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type client_secret: ~azure.synapse.artifacts.models.SecretBase :param access_token: The access token obtained when initially authenticating your OAuth integration. - :type access_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type access_token: ~azure.synapse.artifacts.models.SecretBase :param refresh_token: The refresh token obtained when initially authenticating your OAuth integration. - :type refresh_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type refresh_token: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -17894,16 +18057,14 @@ class HubspotObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -18004,19 +18165,19 @@ class IfConditionActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param expression: Required. An expression that would evaluate to Boolean. This is used to determine the block of activities (ifTrueActivities or ifFalseActivities) that will be executed. - :type expression: ~azure.synapse.artifacts.v2021_06_01_preview.models.Expression + :type expression: ~azure.synapse.artifacts.models.Expression :param if_true_activities: List of activities to execute if expression is evaluated to true. This is an optional property and if not provided, the activity will exit without any action. - :type if_true_activities: list[~azure.synapse.artifacts.v2021_06_01_preview.models.Activity] + :type if_true_activities: list[~azure.synapse.artifacts.models.Activity] :param if_false_activities: List of activities to execute if expression is evaluated to false. This is an optional property and if not provided, the activity will exit without any action. - :type if_false_activities: list[~azure.synapse.artifacts.v2021_06_01_preview.models.Activity] + :type if_false_activities: list[~azure.synapse.artifacts.models.Activity] """ _validation = { @@ -18059,13 +18220,11 @@ class ImpalaLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The IP address or host name of the Impala server. (i.e. @@ -18076,13 +18235,12 @@ class ImpalaLinkedService(LinkedService): :type port: any :param authentication_type: Required. The authentication type to use. Possible values include: "Anonymous", "SASLUsername", "UsernameAndPassword". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.ImpalaAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.ImpalaAuthenticationType :param username: The user name used to access the Impala server. The default value is anonymous when using SASLUsername. :type username: any :param password: The password corresponding to the user name when using UsernameAndPassword. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The default value is false. :type enable_ssl: any @@ -18169,16 +18327,14 @@ class ImpalaObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -18286,13 +18442,11 @@ class InformixLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The non-access credential portion of the connection string @@ -18305,12 +18459,12 @@ class InformixLinkedService(LinkedService): :type authentication_type: any :param credential: The access credential portion of the connection string specified in driver-specific property-value format. - :type credential: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type credential: ~azure.synapse.artifacts.models.SecretBase :param user_name: User name for Basic authentication. Type: string (or Expression with resultType string). :type user_name: any :param password: Password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -18477,16 +18631,14 @@ class InformixTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The Informix table name. Type: string (or Expression with resultType string). :type table_name: any @@ -18532,7 +18684,7 @@ class IntegrationRuntime(msrest.serialization.Model): :type additional_properties: dict[str, any] :param type: Required. Type of integration runtime.Constant filled by server. Possible values include: "Managed", "SelfHosted". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeType + :type type: str or ~azure.synapse.artifacts.models.IntegrationRuntimeType :param description: Integration runtime description. :type description: str """ @@ -18580,10 +18732,9 @@ class IntegrationRuntimeComputeProperties(msrest.serialization.Model): :type max_parallel_executions_per_node: int :param data_flow_properties: Data flow properties for managed integration runtime. :type data_flow_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeDataFlowProperties + ~azure.synapse.artifacts.models.IntegrationRuntimeDataFlowProperties :param v_net_properties: VNet properties for managed integration runtime. - :type v_net_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeVNetProperties + :type v_net_properties: ~azure.synapse.artifacts.models.IntegrationRuntimeVNetProperties """ _validation = { @@ -18622,7 +18773,7 @@ class IntegrationRuntimeCustomSetupScriptProperties(msrest.serialization.Model): script. :type blob_container_uri: str :param sas_token: The SAS token of the Azure blob container. - :type sas_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecureString + :type sas_token: ~azure.synapse.artifacts.models.SecureString """ _attribute_map = { @@ -18647,14 +18798,16 @@ class IntegrationRuntimeDataFlowProperties(msrest.serialization.Model): :type additional_properties: dict[str, any] :param compute_type: Compute type of the cluster which will execute data flow job. Possible values include: "General", "MemoryOptimized", "ComputeOptimized". - :type compute_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowComputeType + :type compute_type: str or ~azure.synapse.artifacts.models.DataFlowComputeType :param core_count: Core count of the cluster which will execute data flow job. Supported values are: 8, 16, 32, 48, 80, 144 and 272. :type core_count: int :param time_to_live: Time to live (in minutes) setting of the cluster which will execute data flow job. :type time_to_live: int + :param cleanup: Cluster will not be recycled and it will be used in next data flow activity run + until TTL (time to live) is reached if this is set as false. Default is true. + :type cleanup: bool """ _validation = { @@ -18666,6 +18819,7 @@ class IntegrationRuntimeDataFlowProperties(msrest.serialization.Model): 'compute_type': {'key': 'computeType', 'type': 'str'}, 'core_count': {'key': 'coreCount', 'type': 'int'}, 'time_to_live': {'key': 'timeToLive', 'type': 'int'}, + 'cleanup': {'key': 'cleanup', 'type': 'bool'}, } def __init__( @@ -18677,16 +18831,16 @@ def __init__( self.compute_type = kwargs.get('compute_type', None) self.core_count = kwargs.get('core_count', None) self.time_to_live = kwargs.get('time_to_live', None) + self.cleanup = kwargs.get('cleanup', None) class IntegrationRuntimeDataProxyProperties(msrest.serialization.Model): """Data proxy properties for a managed dedicated integration runtime. :param connect_via: The self-hosted integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2021_06_01_preview.models.EntityReference + :type connect_via: ~azure.synapse.artifacts.models.EntityReference :param staging_linked_service: The staging linked service reference. - :type staging_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.EntityReference + :type staging_linked_service: ~azure.synapse.artifacts.models.EntityReference :param path: The path to contain the staged data in the Blob storage. :type path: str """ @@ -18707,14 +18861,41 @@ def __init__( self.path = kwargs.get('path', None) +class IntegrationRuntimeDebugResource(SubResourceDebugResource): + """Integration runtime debug resource. + + All required parameters must be populated in order to send to Azure. + + :param name: The resource name. + :type name: str + :param properties: Required. Integration runtime properties. + :type properties: ~azure.synapse.artifacts.models.IntegrationRuntime + """ + + _validation = { + 'properties': {'required': True}, + } + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'properties': {'key': 'properties', 'type': 'IntegrationRuntime'}, + } + + def __init__( + self, + **kwargs + ): + super(IntegrationRuntimeDebugResource, self).__init__(**kwargs) + self.properties = kwargs['properties'] + + class IntegrationRuntimeListResponse(msrest.serialization.Model): """A list of integration runtime resources. All required parameters must be populated in order to send to Azure. :param value: Required. List of integration runtimes. - :type value: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeResource] + :type value: list[~azure.synapse.artifacts.models.IntegrationRuntimeResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -18744,8 +18925,7 @@ class IntegrationRuntimeReference(msrest.serialization.Model): :param type: Required. Type of integration runtime. Possible values include: "IntegrationRuntimeReference". - :type type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReferenceType + :type type: str or ~azure.synapse.artifacts.models.IntegrationRuntimeReferenceType :param reference_name: Required. Reference integration runtime name. :type reference_name: str :param parameters: Arguments for integration runtime. @@ -18791,7 +18971,7 @@ class IntegrationRuntimeResource(SubResource): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Integration runtime properties. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntime + :type properties: ~azure.synapse.artifacts.models.IntegrationRuntime """ _validation = { @@ -18830,12 +19010,12 @@ class IntegrationRuntimeSsisCatalogInfo(msrest.serialization.Model): :type catalog_admin_user_name: str :param catalog_admin_password: The password of the administrator user account of the catalog database. - :type catalog_admin_password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecureString + :type catalog_admin_password: ~azure.synapse.artifacts.models.SecureString :param catalog_pricing_tier: The pricing tier for the catalog database. The valid values could be found in https://azure.microsoft.com/en-us/pricing/details/sql-database/. Possible values include: "Basic", "Standard", "Premium", "PremiumRS". :type catalog_pricing_tier: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeSsisCatalogPricingTier + ~azure.synapse.artifacts.models.IntegrationRuntimeSsisCatalogPricingTier """ _validation = { @@ -18869,28 +19049,24 @@ class IntegrationRuntimeSsisProperties(msrest.serialization.Model): collection. :type additional_properties: dict[str, any] :param catalog_info: Catalog information for managed dedicated integration runtime. - :type catalog_info: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeSsisCatalogInfo + :type catalog_info: ~azure.synapse.artifacts.models.IntegrationRuntimeSsisCatalogInfo :param license_type: License type for bringing your own license scenario. Possible values include: "BasePrice", "LicenseIncluded". - :type license_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeLicenseType + :type license_type: str or ~azure.synapse.artifacts.models.IntegrationRuntimeLicenseType :param custom_setup_script_properties: Custom setup script properties for a managed dedicated integration runtime. :type custom_setup_script_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeCustomSetupScriptProperties + ~azure.synapse.artifacts.models.IntegrationRuntimeCustomSetupScriptProperties :param data_proxy_properties: Data proxy properties for a managed dedicated integration runtime. :type data_proxy_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeDataProxyProperties + ~azure.synapse.artifacts.models.IntegrationRuntimeDataProxyProperties :param edition: The edition for the SSIS Integration Runtime. Possible values include: "Standard", "Enterprise". - :type edition: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeEdition + :type edition: str or ~azure.synapse.artifacts.models.IntegrationRuntimeEdition :param express_custom_setup_properties: Custom setup without script properties for a SSIS integration runtime. - :type express_custom_setup_properties: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.CustomSetupBase] + :type express_custom_setup_properties: list[~azure.synapse.artifacts.models.CustomSetupBase] """ _attribute_map = { @@ -18961,13 +19137,11 @@ class JiraLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The IP address or host name of the Jira service. (e.g. @@ -18980,7 +19154,7 @@ class JiraLinkedService(LinkedService): :type username: any :param password: The password corresponding to the user name that you provided in the username field. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -19055,16 +19229,14 @@ class JiraObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -19169,18 +19341,16 @@ class JsonDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param location: The location of the json data storage. - :type location: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetLocation + :type location: ~azure.synapse.artifacts.models.DatasetLocation :param encoding_name: The code page name of the preferred encoding. If not specified, the default value is UTF-8, unless BOM denotes another Unicode encoding. Refer to the name column of the table in the following link to set supported values: @@ -19188,7 +19358,7 @@ class JsonDataset(Dataset): resultType string). :type encoding_name: any :param compression: The data compression method used for the json dataset. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression """ _validation = { @@ -19237,10 +19407,8 @@ class JsonFormat(DatasetStorageFormat): :param deserializer: Deserializer. Type: string (or Expression with resultType string). :type deserializer: any :param file_pattern: File pattern of JSON. To be more specific, the way of separating a - collection of JSON objects. The default value is 'setOfObjects'. It is case-sensitive. Possible - values include: "setOfObjects", "arrayOfObjects". - :type file_pattern: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.JsonFormatFilePattern + collection of JSON objects. The default value is 'setOfObjects'. It is case-sensitive. + :type file_pattern: any :param nesting_separator: The character used to separate nesting levels. Default value is '.' (dot). Type: string (or Expression with resultType string). :type nesting_separator: any @@ -19270,7 +19438,7 @@ class JsonFormat(DatasetStorageFormat): 'type': {'key': 'type', 'type': 'str'}, 'serializer': {'key': 'serializer', 'type': 'object'}, 'deserializer': {'key': 'deserializer', 'type': 'object'}, - 'file_pattern': {'key': 'filePattern', 'type': 'str'}, + 'file_pattern': {'key': 'filePattern', 'type': 'object'}, 'nesting_separator': {'key': 'nestingSeparator', 'type': 'object'}, 'encoding_name': {'key': 'encodingName', 'type': 'object'}, 'json_node_reference': {'key': 'jsonNodeReference', 'type': 'object'}, @@ -19301,8 +19469,7 @@ class JsonReadSettings(FormatReadSettings): :param type: Required. The read setting type.Constant filled by server. :type type: str :param compression_properties: Compression settings. - :type compression_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.CompressionReadSettings + :type compression_properties: ~azure.synapse.artifacts.models.CompressionReadSettings """ _validation = { @@ -19350,9 +19517,9 @@ class JsonSink(CopySink): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Json store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreWriteSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreWriteSettings :param format_settings: Json format settings. - :type format_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.JsonWriteSettings + :type format_settings: ~azure.synapse.artifacts.models.JsonWriteSettings """ _validation = { @@ -19401,9 +19568,9 @@ class JsonSource(CopySource): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Json store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings :param format_settings: Json format settings. - :type format_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.JsonReadSettings + :type format_settings: ~azure.synapse.artifacts.models.JsonReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :type additional_columns: any @@ -19446,10 +19613,8 @@ class JsonWriteSettings(FormatWriteSettings): :param type: Required. The write setting type.Constant filled by server. :type type: str :param file_pattern: File pattern of JSON. This setting controls the way a collection of JSON - objects will be treated. The default value is 'setOfObjects'. It is case-sensitive. Possible - values include: "setOfObjects", "arrayOfObjects". - :type file_pattern: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.JsonWriteFilePattern + objects will be treated. The default value is 'setOfObjects'. It is case-sensitive. + :type file_pattern: any """ _validation = { @@ -19459,7 +19624,7 @@ class JsonWriteSettings(FormatWriteSettings): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, 'type': {'key': 'type', 'type': 'str'}, - 'file_pattern': {'key': 'filePattern', 'type': 'str'}, + 'file_pattern': {'key': 'filePattern', 'type': 'object'}, } def __init__( @@ -19471,6 +19636,148 @@ def __init__( self.file_pattern = kwargs.get('file_pattern', None) +class KqlScript(msrest.serialization.Model): + """KqlScript. + + :param content: + :type content: ~azure.synapse.artifacts.models.KqlScriptContent + """ + + _attribute_map = { + 'content': {'key': 'content', 'type': 'KqlScriptContent'}, + } + + def __init__( + self, + **kwargs + ): + super(KqlScript, self).__init__(**kwargs) + self.content = kwargs.get('content', None) + + +class KqlScriptContent(msrest.serialization.Model): + """KqlScriptContent. + + :param query: + :type query: str + :param metadata: + :type metadata: ~azure.synapse.artifacts.models.KqlScriptContentMetadata + :param current_connection: + :type current_connection: ~azure.synapse.artifacts.models.KqlScriptContentCurrentConnection + """ + + _attribute_map = { + 'query': {'key': 'query', 'type': 'str'}, + 'metadata': {'key': 'metadata', 'type': 'KqlScriptContentMetadata'}, + 'current_connection': {'key': 'currentConnection', 'type': 'KqlScriptContentCurrentConnection'}, + } + + def __init__( + self, + **kwargs + ): + super(KqlScriptContent, self).__init__(**kwargs) + self.query = kwargs.get('query', None) + self.metadata = kwargs.get('metadata', None) + self.current_connection = kwargs.get('current_connection', None) + + +class KqlScriptContentCurrentConnection(msrest.serialization.Model): + """KqlScriptContentCurrentConnection. + + :param name: + :type name: str + :param type: + :type type: str + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__( + self, + **kwargs + ): + super(KqlScriptContentCurrentConnection, self).__init__(**kwargs) + self.name = kwargs.get('name', None) + self.type = kwargs.get('type', None) + + +class KqlScriptContentMetadata(msrest.serialization.Model): + """KqlScriptContentMetadata. + + :param language: + :type language: str + """ + + _attribute_map = { + 'language': {'key': 'language', 'type': 'str'}, + } + + def __init__( + self, + **kwargs + ): + super(KqlScriptContentMetadata, self).__init__(**kwargs) + self.language = kwargs.get('language', None) + + +class KqlScriptResource(msrest.serialization.Model): + """KqlScriptResource. + + :param id: + :type id: str + :param name: + :type name: str + :param type: + :type type: str + :param properties: Properties of sql script. + :type properties: ~azure.synapse.artifacts.models.KqlScript + """ + + _attribute_map = { + 'id': {'key': 'id', 'type': 'str'}, + 'name': {'key': 'name', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'properties': {'key': 'properties', 'type': 'KqlScript'}, + } + + def __init__( + self, + **kwargs + ): + super(KqlScriptResource, self).__init__(**kwargs) + self.id = kwargs.get('id', None) + self.name = kwargs.get('name', None) + self.type = kwargs.get('type', None) + self.properties = kwargs.get('properties', None) + + +class KqlScriptsResourceCollectionResponse(msrest.serialization.Model): + """KqlScriptsResourceCollectionResponse. + + :param value: + :type value: list[~azure.synapse.artifacts.models.KqlScriptResource] + :param next_link: + :type next_link: str + """ + + _attribute_map = { + 'value': {'key': 'value', 'type': '[KqlScriptResource]'}, + 'next_link': {'key': 'nextLink', 'type': 'str'}, + } + + def __init__( + self, + **kwargs + ): + super(KqlScriptsResourceCollectionResponse, self).__init__(**kwargs) + self.value = kwargs.get('value', None) + self.next_link = kwargs.get('next_link', None) + + class LibraryInfo(msrest.serialization.Model): """Library/package information of a Big Data pool powered by Apache Spark. @@ -19482,8 +19789,8 @@ class LibraryInfo(msrest.serialization.Model): :type path: str :param container_name: Storage blob container name. :type container_name: str - :ivar uploaded_timestamp: The last update time of the library. - :vartype uploaded_timestamp: ~datetime.datetime + :param uploaded_timestamp: The last update time of the library. + :type uploaded_timestamp: ~datetime.datetime :param type: Type of the library. :type type: str :ivar provisioning_status: Provisioning status of the library/package. @@ -19493,7 +19800,6 @@ class LibraryInfo(msrest.serialization.Model): """ _validation = { - 'uploaded_timestamp': {'readonly': True}, 'provisioning_status': {'readonly': True}, 'creator_id': {'readonly': True}, } @@ -19516,7 +19822,7 @@ def __init__( self.name = kwargs.get('name', None) self.path = kwargs.get('path', None) self.container_name = kwargs.get('container_name', None) - self.uploaded_timestamp = None + self.uploaded_timestamp = kwargs.get('uploaded_timestamp', None) self.type = kwargs.get('type', None) self.provisioning_status = None self.creator_id = None @@ -19528,7 +19834,7 @@ class LibraryListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of Library. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.LibraryResource] + :type value: list[~azure.synapse.artifacts.models.LibraryResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -19602,7 +19908,7 @@ class LibraryResource(SubResource): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Library/package properties. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.LibraryResourceProperties + :type properties: ~azure.synapse.artifacts.models.LibraryResourceProperties """ _validation = { @@ -19792,7 +20098,7 @@ class LinkedIntegrationRuntimeKeyAuthorization(LinkedIntegrationRuntimeType): sharing.Constant filled by server. :type authorization_type: str :param key: Required. The key used for authorization. - :type key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecureString + :type key: ~azure.synapse.artifacts.models.SecureString """ _validation = { @@ -19853,7 +20159,7 @@ class LinkedServiceDebugResource(SubResourceDebugResource): :param name: The resource name. :type name: str :param properties: Required. Properties of linked service. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedService + :type properties: ~azure.synapse.artifacts.models.LinkedService """ _validation = { @@ -19879,7 +20185,7 @@ class LinkedServiceListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of linked services. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceResource] + :type value: list[~azure.synapse.artifacts.models.LinkedServiceResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -19909,7 +20215,7 @@ class LinkedServiceReference(msrest.serialization.Model): :param type: Required. Linked service reference type. Possible values include: "LinkedServiceReference". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.Type + :type type: str or ~azure.synapse.artifacts.models.Type :param reference_name: Required. Reference LinkedService name. :type reference_name: str :param parameters: Arguments for LinkedService. @@ -19955,7 +20261,7 @@ class LinkedServiceResource(SubResource): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Properties of linked service. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedService + :type properties: ~azure.synapse.artifacts.models.LinkedService """ _validation = { @@ -19988,8 +20294,7 @@ class LogLocationSettings(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param linked_service_name: Required. Log storage linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param path: The path to storage for storing detailed logs of activity execution. Type: string (or Expression with resultType string). :type path: any @@ -20022,12 +20327,10 @@ class LogSettings(msrest.serialization.Model): (or Expression with resultType boolean). :type enable_copy_activity_log: any :param copy_activity_log_settings: Specifies settings for copy activity log. - :type copy_activity_log_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.CopyActivityLogSettings + :type copy_activity_log_settings: ~azure.synapse.artifacts.models.CopyActivityLogSettings :param log_location_settings: Required. Log location settings customer needs to provide when enabling log. - :type log_location_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LogLocationSettings + :type log_location_settings: ~azure.synapse.artifacts.models.LogLocationSettings """ _validation = { @@ -20059,8 +20362,7 @@ class LogStorageSettings(msrest.serialization.Model): collection. :type additional_properties: dict[str, any] :param linked_service_name: Required. Log storage linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param path: The path to storage for storing detailed logs of activity execution. Type: string (or Expression with resultType string). :type path: any @@ -20111,18 +20413,17 @@ class LookupActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param source: Required. Dataset-specific source properties, same as copy activity source. - :type source: ~azure.synapse.artifacts.v2021_06_01_preview.models.CopySource + :type source: ~azure.synapse.artifacts.models.CopySource :param dataset: Required. Lookup activity dataset reference. - :type dataset: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference + :type dataset: ~azure.synapse.artifacts.models.DatasetReference :param first_row_only: Whether to return first row or all rows. Default value is true. Type: boolean (or Expression with resultType boolean). :type first_row_only: any @@ -20171,19 +20472,17 @@ class MagentoLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The URL of the Magento instance. (i.e. 192.168.222.110/magento3). :type host: any :param access_token: The access token from Magento. - :type access_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type access_token: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -20253,16 +20552,14 @@ class MagentoObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -20359,7 +20656,7 @@ class ManagedIdentity(msrest.serialization.Model): :vartype tenant_id: str :param type: The type of managed identity for the workspace. Possible values include: "None", "SystemAssigned". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.ResourceIdentityType + :type type: str or ~azure.synapse.artifacts.models.ResourceIdentityType """ _validation = { @@ -20395,23 +20692,19 @@ class ManagedIntegrationRuntime(IntegrationRuntime): :type additional_properties: dict[str, any] :param type: Required. Type of integration runtime.Constant filled by server. Possible values include: "Managed", "SelfHosted". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeType + :type type: str or ~azure.synapse.artifacts.models.IntegrationRuntimeType :param description: Integration runtime description. :type description: str :ivar state: Integration runtime state, only valid for managed dedicated integration runtime. Possible values include: "Initial", "Stopped", "Started", "Starting", "Stopping", "NeedRegistration", "Online", "Limited", "Offline", "AccessDenied". - :vartype state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeState + :vartype state: str or ~azure.synapse.artifacts.models.IntegrationRuntimeState :param managed_virtual_network: Managed Virtual Network reference. - :type managed_virtual_network: - ~azure.synapse.artifacts.v2021_06_01_preview.models.ManagedVirtualNetworkReference + :type managed_virtual_network: ~azure.synapse.artifacts.models.ManagedVirtualNetworkReference :param compute_properties: The compute resource for managed integration runtime. - :type compute_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeComputeProperties + :type compute_properties: ~azure.synapse.artifacts.models.IntegrationRuntimeComputeProperties :param ssis_properties: SSIS properties for managed integration runtime. - :type ssis_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeSsisProperties + :type ssis_properties: ~azure.synapse.artifacts.models.IntegrationRuntimeSsisProperties """ _validation = { @@ -20515,13 +20808,13 @@ class MappingDataFlow(DataFlow): :type annotations: list[any] :param folder: The folder that this data flow is in. If not specified, Data flow will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowFolder + :type folder: ~azure.synapse.artifacts.models.DataFlowFolder :param sources: List of sources in data flow. - :type sources: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowSource] + :type sources: list[~azure.synapse.artifacts.models.DataFlowSource] :param sinks: List of sinks in data flow. - :type sinks: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowSink] + :type sinks: list[~azure.synapse.artifacts.models.DataFlowSink] :param transformations: List of transformations in data flow. - :type transformations: list[~azure.synapse.artifacts.v2021_06_01_preview.models.Transformation] + :type transformations: list[~azure.synapse.artifacts.models.Transformation] :param script: DataFlow script. :type script: str """ @@ -20564,20 +20857,18 @@ class MariaDBLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type pwd: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -20684,16 +20975,14 @@ class MariaDBTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -20736,13 +21025,11 @@ class MarketoLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param endpoint: Required. The endpoint of the Marketo server. (i.e. 123-ABC-321.mktorest.com). @@ -20750,7 +21037,7 @@ class MarketoLinkedService(LinkedService): :param client_id: Required. The client Id of your Marketo service. :type client_id: any :param client_secret: The client secret of your Marketo service. - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type client_secret: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -20823,16 +21110,14 @@ class MarketoObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -20929,13 +21214,11 @@ class MicrosoftAccessLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The non-access credential portion of the connection string @@ -20948,12 +21231,12 @@ class MicrosoftAccessLinkedService(LinkedService): :type authentication_type: any :param credential: The access credential portion of the connection string specified in driver-specific property-value format. - :type credential: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type credential: ~azure.synapse.artifacts.models.SecretBase :param user_name: User name for Basic authentication. Type: string (or Expression with resultType string). :type user_name: any :param password: Password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -21117,16 +21400,14 @@ class MicrosoftAccessTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The Microsoft Access table name. Type: string (or Expression with resultType string). :type table_name: any @@ -21178,16 +21459,14 @@ class MongoDbAtlasCollectionDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param collection: Required. The collection name of the MongoDB Atlas database. Type: string (or Expression with resultType string). :type collection: any @@ -21232,13 +21511,11 @@ class MongoDbAtlasLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The MongoDB Atlas connection string. Type: string, @@ -21301,8 +21578,7 @@ class MongoDbAtlasSource(CopySource): with resultType string). :type filter: any :param cursor_methods: Cursor methods for Mongodb query. - :type cursor_methods: - ~azure.synapse.artifacts.v2021_06_01_preview.models.MongoDbCursorMethodsProperties + :type cursor_methods: ~azure.synapse.artifacts.models.MongoDbCursorMethodsProperties :param batch_size: Specifies the number of documents to return in each batch of the response from MongoDB Atlas instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response @@ -21365,16 +21641,14 @@ class MongoDbCollectionDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param collection_name: Required. The table name of the MongoDB database. Type: string (or Expression with resultType string). :type collection_name: any @@ -21462,13 +21736,11 @@ class MongoDbLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param server: Required. The IP address or server name of the MongoDB server. Type: string (or @@ -21476,8 +21748,7 @@ class MongoDbLinkedService(LinkedService): :type server: any :param authentication_type: The authentication type to be used to connect to the MongoDB database. Possible values include: "Basic", "Anonymous". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.MongoDbAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.MongoDbAuthenticationType :param database_name: Required. The name of the MongoDB database that you want to access. Type: string (or Expression with resultType string). :type database_name: any @@ -21485,7 +21756,7 @@ class MongoDbLinkedService(LinkedService): string). :type username: any :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param auth_source: Database to verify the username and password. Type: string (or Expression with resultType string). :type auth_source: any @@ -21617,16 +21888,14 @@ class MongoDbV2CollectionDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param collection: Required. The collection name of the MongoDB database. Type: string (or Expression with resultType string). :type collection: any @@ -21671,13 +21940,11 @@ class MongoDbV2LinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The MongoDB connection string. Type: string, SecureString @@ -21739,8 +22006,7 @@ class MongoDbV2Source(CopySource): with resultType string). :type filter: any :param cursor_methods: Cursor methods for Mongodb query. - :type cursor_methods: - ~azure.synapse.artifacts.v2021_06_01_preview.models.MongoDbCursorMethodsProperties + :type cursor_methods: ~azure.synapse.artifacts.models.MongoDbCursorMethodsProperties :param batch_size: Specifies the number of documents to return in each batch of the response from MongoDB instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. @@ -21795,20 +22061,17 @@ class MySqlLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The connection string. :type connection_string: any :param password: The Azure key vault secret reference of password in connection string. - :type password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type password: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -21915,16 +22178,14 @@ class MySqlTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The MySQL table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -21967,20 +22228,18 @@ class NetezzaLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type pwd: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -22076,11 +22335,9 @@ class NetezzaSource(TabularSource): :type query: any :param partition_option: The partition mechanism that will be used for Netezza read in parallel. Possible values include: "None", "DataSlice", "DynamicRange". - :type partition_option: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.NetezzaPartitionOption + :type partition_option: str or ~azure.synapse.artifacts.models.NetezzaPartitionOption :param partition_settings: The settings that will be leveraged for Netezza source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.NetezzaPartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.NetezzaPartitionSettings """ _validation = { @@ -22130,16 +22387,14 @@ class NetezzaTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -22193,12 +22448,11 @@ class Notebook(msrest.serialization.Model): :param description: The description of the notebook. :type description: str :param big_data_pool: Big data pool reference. - :type big_data_pool: ~azure.synapse.artifacts.v2021_06_01_preview.models.BigDataPoolReference + :type big_data_pool: ~azure.synapse.artifacts.models.BigDataPoolReference :param session_properties: Session properties. - :type session_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookSessionProperties + :type session_properties: ~azure.synapse.artifacts.models.NotebookSessionProperties :param metadata: Required. Notebook root-level metadata. - :type metadata: ~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookMetadata + :type metadata: ~azure.synapse.artifacts.models.NotebookMetadata :param nbformat: Required. Notebook format (major number). Incremented between backwards incompatible changes to the notebook format. :type nbformat: int @@ -22206,7 +22460,10 @@ class Notebook(msrest.serialization.Model): compatible changes to the notebook format. :type nbformat_minor: int :param cells: Required. Array of cells of the current notebook. - :type cells: list[~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookCell] + :type cells: list[~azure.synapse.artifacts.models.NotebookCell] + :param folder: The folder that this notebook is in. If not specified, this notebook will appear + at the root level. + :type folder: ~azure.synapse.artifacts.models.NotebookFolder """ _validation = { @@ -22225,6 +22482,7 @@ class Notebook(msrest.serialization.Model): 'nbformat': {'key': 'nbformat', 'type': 'int'}, 'nbformat_minor': {'key': 'nbformat_minor', 'type': 'int'}, 'cells': {'key': 'cells', 'type': '[NotebookCell]'}, + 'folder': {'key': 'folder', 'type': 'NotebookFolder'}, } def __init__( @@ -22240,6 +22498,7 @@ def __init__( self.nbformat = kwargs['nbformat'] self.nbformat_minor = kwargs['nbformat_minor'] self.cells = kwargs['cells'] + self.folder = kwargs.get('folder', None) class NotebookCell(msrest.serialization.Model): @@ -22259,7 +22518,7 @@ class NotebookCell(msrest.serialization.Model): :param attachments: Attachments associated with the cell. :type attachments: any :param outputs: Cell-level output items. - :type outputs: list[~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookCellOutputItem] + :type outputs: list[~azure.synapse.artifacts.models.NotebookCellOutputItem] """ _validation = { @@ -22301,7 +22560,7 @@ class NotebookCellOutputItem(msrest.serialization.Model): :type execution_count: int :param output_type: Required. Execution, display, or stream outputs. Possible values include: "execute_result", "display_data", "stream", "error". - :type output_type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.CellOutputType + :type output_type: str or ~azure.synapse.artifacts.models.CellOutputType :param text: For output_type=stream, the stream's text output, represented as a string or an array of strings. :type text: any @@ -22337,6 +22596,25 @@ def __init__( self.metadata = kwargs.get('metadata', None) +class NotebookFolder(msrest.serialization.Model): + """The folder that this notebook is in. If not specified, this notebook will appear at the root level. + + :param name: The name of the folder that this notebook is in. + :type name: str + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + } + + def __init__( + self, + **kwargs + ): + super(NotebookFolder, self).__init__(**kwargs) + self.name = kwargs.get('name', None) + + class NotebookKernelSpec(msrest.serialization.Model): """Kernel information. @@ -22412,7 +22690,7 @@ class NotebookListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of Notebooks. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookResource] + :type value: list[~azure.synapse.artifacts.models.NotebookResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -22442,9 +22720,9 @@ class NotebookMetadata(msrest.serialization.Model): collection. :type additional_properties: dict[str, any] :param kernelspec: Kernel information. - :type kernelspec: ~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookKernelSpec + :type kernelspec: ~azure.synapse.artifacts.models.NotebookKernelSpec :param language_info: Language info. - :type language_info: ~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookLanguageInfo + :type language_info: ~azure.synapse.artifacts.models.NotebookLanguageInfo """ _attribute_map = { @@ -22481,7 +22759,7 @@ class NotebookResource(msrest.serialization.Model): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Properties of Notebook. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.Notebook + :type properties: ~azure.synapse.artifacts.models.Notebook """ _validation = { @@ -22568,13 +22846,11 @@ class ODataLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param url: Required. The URL of the OData service endpoint. Type: string (or Expression with @@ -22583,13 +22859,12 @@ class ODataLinkedService(LinkedService): :param authentication_type: Type of authentication used to connect to the OData service. Possible values include: "Basic", "Anonymous", "Windows", "AadServicePrincipal", "ManagedServiceIdentity". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.ODataAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.ODataAuthenticationType :param user_name: User name of the OData service. Type: string (or Expression with resultType string). :type user_name: any :param password: Password of the OData service. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param tenant: Specify the tenant information (domain name or tenant ID) under which your application resides. Type: string (or Expression with resultType string). :type tenant: any @@ -22606,20 +22881,18 @@ class ODataLinkedService(LinkedService): :param aad_service_principal_credential_type: Specify the credential type (key or cert) is used for service principal. Possible values include: "ServicePrincipalKey", "ServicePrincipalCert". :type aad_service_principal_credential_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.ODataAadServicePrincipalCredentialType + ~azure.synapse.artifacts.models.ODataAadServicePrincipalCredentialType :param service_principal_key: Specify the secret of your application registered in Azure Active Directory. Type: string (or Expression with resultType string). - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param service_principal_embedded_cert: Specify the base64 encoded certificate of your application registered in Azure Active Directory. Type: string (or Expression with resultType string). - :type service_principal_embedded_cert: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_embedded_cert: ~azure.synapse.artifacts.models.SecretBase :param service_principal_embedded_cert_password: Specify the password of your certificate if your certificate has a password and you are using AadServicePrincipal authentication. Type: string (or Expression with resultType string). - :type service_principal_embedded_cert_password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_embedded_cert_password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -22693,16 +22966,14 @@ class ODataResourceDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param path: The OData resource path. Type: string (or Expression with resultType string). :type path: any """ @@ -22803,13 +23074,11 @@ class OdbcLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The non-access credential portion of the connection string @@ -22821,12 +23090,12 @@ class OdbcLinkedService(LinkedService): :type authentication_type: any :param credential: The access credential portion of the connection string specified in driver-specific property-value format. - :type credential: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type credential: ~azure.synapse.artifacts.models.SecretBase :param user_name: User name for Basic authentication. Type: string (or Expression with resultType string). :type user_name: any :param password: Password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -22993,16 +23262,14 @@ class OdbcTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The ODBC table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -23053,16 +23320,14 @@ class Office365Dataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: Required. Name of the dataset to extract from Office 365. Type: string (or Expression with resultType string). :type table_name: any @@ -23112,13 +23377,11 @@ class Office365LinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param office365_tenant_id: Required. Azure tenant ID to which the Office 365 account belongs. @@ -23131,7 +23394,7 @@ class Office365LinkedService(LinkedService): Expression with resultType string). :type service_principal_id: any :param service_principal_key: Required. Specify the application's key. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -23259,7 +23522,7 @@ class OperationResult(msrest.serialization.Model): :param target: Property name/path in request associated with error. :type target: str :param details: Array with additional error details. - :type details: list[~azure.synapse.artifacts.v2021_06_01_preview.models.CloudError] + :type details: list[~azure.synapse.artifacts.models.CloudErrorAutoGenerated] """ _validation = { @@ -23271,7 +23534,7 @@ class OperationResult(msrest.serialization.Model): 'code': {'key': 'error.code', 'type': 'str'}, 'message': {'key': 'error.message', 'type': 'str'}, 'target': {'key': 'error.target', 'type': 'str'}, - 'details': {'key': 'error.details', 'type': '[CloudError]'}, + 'details': {'key': 'error.details', 'type': '[CloudErrorAutoGenerated]'}, } def __init__( @@ -23297,21 +23560,18 @@ class OracleLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param password: The Azure key vault secret reference of password in connection string. - :type password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type password: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -23393,13 +23653,11 @@ class OracleServiceCloudLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The URL of the Oracle Service Cloud instance. @@ -23408,7 +23666,7 @@ class OracleServiceCloudLinkedService(LinkedService): :type username: any :param password: Required. The password corresponding to the user name that you provided in the username key. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. Type: boolean (or Expression with resultType boolean). :type use_encrypted_endpoints: any @@ -23483,16 +23741,14 @@ class OracleServiceCloudObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -23659,11 +23915,9 @@ class OracleSource(CopySource): :type query_timeout: any :param partition_option: The partition mechanism that will be used for Oracle read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". - :type partition_option: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.OraclePartitionOption + :type partition_option: str or ~azure.synapse.artifacts.models.OraclePartitionOption :param partition_settings: The settings that will be leveraged for Oracle source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.OraclePartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.OraclePartitionSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :type additional_columns: any @@ -23718,16 +23972,14 @@ class OracleTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -23789,21 +24041,19 @@ class OrcDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param location: The location of the ORC data storage. - :type location: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetLocation - :param orc_compression_codec: Possible values include: "none", "zlib", "snappy", "lzo". - :type orc_compression_codec: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.OrcCompressionCodec + :type location: ~azure.synapse.artifacts.models.DatasetLocation + :param orc_compression_codec: The data orcCompressionCodec. Type: string (or Expression with + resultType string). + :type orc_compression_codec: any """ _validation = { @@ -23822,7 +24072,7 @@ class OrcDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'location': {'key': 'typeProperties.location', 'type': 'DatasetLocation'}, - 'orc_compression_codec': {'key': 'typeProperties.orcCompressionCodec', 'type': 'str'}, + 'orc_compression_codec': {'key': 'typeProperties.orcCompressionCodec', 'type': 'object'}, } def __init__( @@ -23896,9 +24146,9 @@ class OrcSink(CopySink): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: ORC store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreWriteSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreWriteSettings :param format_settings: ORC format settings. - :type format_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.OrcWriteSettings + :type format_settings: ~azure.synapse.artifacts.models.OrcWriteSettings """ _validation = { @@ -23947,7 +24197,7 @@ class OrcSource(CopySource): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: ORC store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :type additional_columns: any @@ -24024,7 +24274,7 @@ class ParameterSpecification(msrest.serialization.Model): :param type: Required. Parameter type. Possible values include: "Object", "String", "Int", "Float", "Bool", "Array", "SecureString". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterType + :type type: str or ~azure.synapse.artifacts.models.ParameterType :param default_value: Default value of parameter. :type default_value: any """ @@ -24066,18 +24316,16 @@ class ParquetDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param location: The location of the parquet storage. - :type location: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetLocation + :type location: ~azure.synapse.artifacts.models.DatasetLocation :param compression_codec: A string from ParquetCompressionCodecEnum or an expression. :type compression_codec: any """ @@ -24172,9 +24420,9 @@ class ParquetSink(CopySink): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Parquet store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreWriteSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreWriteSettings :param format_settings: Parquet format settings. - :type format_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.ParquetWriteSettings + :type format_settings: ~azure.synapse.artifacts.models.ParquetWriteSettings """ _validation = { @@ -24223,7 +24471,7 @@ class ParquetSource(CopySource): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Parquet store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :type additional_columns: any @@ -24304,13 +24552,11 @@ class PaypalLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The URL of the PayPal instance. (i.e. api.sandbox.paypal.com). @@ -24318,7 +24564,7 @@ class PaypalLinkedService(LinkedService): :param client_id: Required. The client ID associated with your PayPal application. :type client_id: any :param client_secret: The client secret associated with your PayPal application. - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type client_secret: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -24391,16 +24637,14 @@ class PaypalObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -24497,13 +24741,11 @@ class PhoenixLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The IP address or host name of the Phoenix server. (i.e. @@ -24519,12 +24761,11 @@ class PhoenixLinkedService(LinkedService): :param authentication_type: Required. The authentication mechanism used to connect to the Phoenix server. Possible values include: "Anonymous", "UsernameAndPassword", "WindowsAzureHDInsightService". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.PhoenixAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.PhoenixAuthenticationType :param username: The user name used to connect to the Phoenix server. :type username: any :param password: The password corresponding to the user name. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The default value is false. :type enable_ssl: any @@ -24613,16 +24854,14 @@ class PhoenixObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -24744,7 +24983,7 @@ class PipelineListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of pipelines. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineResource] + :type value: list[~azure.synapse.artifacts.models.PipelineResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -24773,7 +25012,7 @@ class PipelineReference(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param type: Required. Pipeline reference type. Possible values include: "PipelineReference". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineReferenceType + :type type: str or ~azure.synapse.artifacts.models.PipelineReferenceType :param reference_name: Required. Reference pipeline name. :type reference_name: str :param name: Reference name. @@ -24822,13 +25061,11 @@ class PipelineResource(SubResource): :param description: The description of the pipeline. :type description: str :param activities: List of activities in pipeline. - :type activities: list[~azure.synapse.artifacts.v2021_06_01_preview.models.Activity] + :type activities: list[~azure.synapse.artifacts.models.Activity] :param parameters: List of parameters for pipeline. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param variables: List of variables for pipeline. - :type variables: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.VariableSpecification] + :type variables: dict[str, ~azure.synapse.artifacts.models.VariableSpecification] :param concurrency: The max number of concurrent runs for the pipeline. :type concurrency: int :param annotations: List of tags that can be used for describing the Pipeline. @@ -24837,7 +25074,7 @@ class PipelineResource(SubResource): :type run_dimensions: dict[str, any] :param folder: The folder that this Pipeline is in. If not specified, Pipeline will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineFolder + :type folder: ~azure.synapse.artifacts.models.PipelineFolder """ _validation = { @@ -24900,7 +25137,7 @@ class PipelineRun(msrest.serialization.Model): run. :vartype parameters: dict[str, str] :ivar invoked_by: Entity that started the pipeline run. - :vartype invoked_by: ~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineRunInvokedBy + :vartype invoked_by: ~azure.synapse.artifacts.models.PipelineRunInvokedBy :ivar last_updated: The last updated timestamp for the pipeline run event in ISO8601 format. :vartype last_updated: ~datetime.datetime :ivar run_start: The start time of a pipeline run in ISO8601 format. @@ -25007,7 +25244,7 @@ class PipelineRunsQueryResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of pipeline runs. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineRun] + :type value: list[~azure.synapse.artifacts.models.PipelineRun] :param continuation_token: The continuation token for getting the next page of results, if any remaining results exist, null otherwise. :type continuation_token: str @@ -25038,8 +25275,7 @@ class PolybaseSettings(msrest.serialization.Model): collection. :type additional_properties: dict[str, any] :param reject_type: Reject type. Possible values include: "value", "percentage". - :type reject_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.PolybaseSettingsRejectType + :type reject_type: str or ~azure.synapse.artifacts.models.PolybaseSettingsRejectType :param reject_value: Specifies the value or the percentage of rows that can be rejected before the query fails. Type: number (or Expression with resultType number), minimum: 0. :type reject_value: any @@ -25084,20 +25320,17 @@ class PostgreSqlLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The connection string. :type connection_string: any :param password: The Azure key vault secret reference of password in connection string. - :type password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type password: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -25204,16 +25437,14 @@ class PostgreSqlTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -25266,13 +25497,11 @@ class PrestoLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The IP address or host name of the Presto server. (i.e. @@ -25287,12 +25516,11 @@ class PrestoLinkedService(LinkedService): :type port: any :param authentication_type: Required. The authentication mechanism used to connect to the Presto server. Possible values include: "Anonymous", "LDAP". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.PrestoAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.PrestoAuthenticationType :param username: The user name used to connect to the Presto server. :type username: any :param password: The password corresponding to the user name. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The default value is false. :type enable_ssl: any @@ -25390,16 +25618,14 @@ class PrestoObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -25569,11 +25795,11 @@ class PrivateEndpointConnection(ProxyResource): "Microsoft.Storage/storageAccounts". :vartype type: str :param private_endpoint: The private endpoint which the connection belongs to. - :type private_endpoint: ~azure.synapse.artifacts.v2021_06_01_preview.models.PrivateEndpoint + :type private_endpoint: ~azure.synapse.artifacts.models.PrivateEndpoint :param private_link_service_connection_state: Connection state of the private endpoint connection. :type private_link_service_connection_state: - ~azure.synapse.artifacts.v2021_06_01_preview.models.PrivateLinkServiceConnectionState + ~azure.synapse.artifacts.models.PrivateLinkServiceConnectionState :ivar provisioning_state: Provisioning state of the private endpoint connection. :vartype provisioning_state: str """ @@ -25660,7 +25886,7 @@ class QueryDataFlowDebugSessionsResponse(msrest.serialization.Model): """A list of active debug sessions. :param value: Array with all active debug sessions. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowDebugSessionInfo] + :type value: list[~azure.synapse.artifacts.models.DataFlowDebugSessionInfo] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -25690,13 +25916,11 @@ class QuickBooksLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_properties: Properties used to connect to QuickBooks. It is mutually @@ -25710,11 +25934,11 @@ class QuickBooksLinkedService(LinkedService): :param consumer_key: Required. The consumer key for OAuth 1.0 authentication. :type consumer_key: any :param consumer_secret: Required. The consumer secret for OAuth 1.0 authentication. - :type consumer_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type consumer_secret: ~azure.synapse.artifacts.models.SecretBase :param access_token: Required. The access token for OAuth 1.0 authentication. - :type access_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type access_token: ~azure.synapse.artifacts.models.SecretBase :param access_token_secret: Required. The access token secret for OAuth 1.0 authentication. - :type access_token_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type access_token_secret: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -25788,16 +26012,14 @@ class QuickBooksObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -25894,12 +26116,11 @@ class RecurrenceSchedule(msrest.serialization.Model): :param hours: The hours. :type hours: list[int] :param week_days: The days of the week. - :type week_days: list[str or ~azure.synapse.artifacts.v2021_06_01_preview.models.DayOfWeek] + :type week_days: list[str or ~azure.synapse.artifacts.models.DayOfWeek] :param month_days: The month days. :type month_days: list[int] :param monthly_occurrences: The monthly occurrences. - :type monthly_occurrences: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.RecurrenceScheduleOccurrence] + :type monthly_occurrences: list[~azure.synapse.artifacts.models.RecurrenceScheduleOccurrence] """ _attribute_map = { @@ -25932,7 +26153,7 @@ class RecurrenceScheduleOccurrence(msrest.serialization.Model): :type additional_properties: dict[str, any] :param day: The day of the week. Possible values include: "Sunday", "Monday", "Tuesday", "Wednesday", "Thursday", "Friday", "Saturday". - :type day: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.DayOfWeek + :type day: str or ~azure.synapse.artifacts.models.DayOfWeek :param occurrence: The occurrence. :type occurrence: int """ @@ -25998,8 +26219,7 @@ class RedshiftUnloadSettings(msrest.serialization.Model): :param s3_linked_service_name: Required. The name of the Amazon S3 linked service which will be used for the unload operation when copying from the Amazon Redshift source. - :type s3_linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type s3_linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param bucket_name: Required. The bucket of the interim Amazon S3 which will be used to store the unloaded data from Amazon Redshift source. The bucket must be in the same region as the Amazon Redshift source. Type: string (or Expression with resultType string). @@ -26094,16 +26314,14 @@ class RelationalTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The relational table name. Type: string (or Expression with resultType string). :type table_name: any @@ -26144,7 +26362,7 @@ class RerunTriggerListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of rerun triggers. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.RerunTriggerResource] + :type value: list[~azure.synapse.artifacts.models.RerunTriggerResource] :ivar next_link: The continuation token for getting the next page of results, if any remaining results exist, null otherwise. :vartype next_link: str @@ -26187,8 +26405,7 @@ class RerunTriggerResource(SubResource): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Properties of the rerun trigger. - :type properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.RerunTumblingWindowTrigger + :type properties: ~azure.synapse.artifacts.models.RerunTumblingWindowTrigger """ _validation = { @@ -26231,8 +26448,7 @@ class RerunTumblingWindowTrigger(Trigger): :type description: str :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRuntimeState + :vartype runtime_state: str or ~azure.synapse.artifacts.models.TriggerRuntimeState :param annotations: List of tags that can be used for describing the trigger. :type annotations: list[any] :param parent_trigger: Required. The parent trigger reference. @@ -26330,13 +26546,11 @@ class ResponsysLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param endpoint: Required. The endpoint of the Responsys server. @@ -26346,7 +26560,7 @@ class ResponsysLinkedService(LinkedService): :type client_id: any :param client_secret: The client secret associated with the Responsys application. Type: string (or Expression with resultType string). - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type client_secret: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. Type: boolean (or Expression with resultType boolean). :type use_encrypted_endpoints: any @@ -26420,16 +26634,14 @@ class ResponsysObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -26534,16 +26746,14 @@ class RestResourceDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param relative_url: The relative URL to the resource that the RESTful API provides. Type: string (or Expression with resultType string). :type relative_url: any @@ -26607,13 +26817,11 @@ class RestServiceLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param url: Required. The base URL of the REST service. @@ -26625,18 +26833,17 @@ class RestServiceLinkedService(LinkedService): :param authentication_type: Required. Type of authentication used to connect to the REST service. Possible values include: "Anonymous", "Basic", "AadServicePrincipal", "ManagedServiceIdentity". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.RestServiceAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.RestServiceAuthenticationType :param user_name: The user name used in Basic authentication type. :type user_name: any :param password: The password used in Basic authentication type. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param service_principal_id: The application's client ID used in AadServicePrincipal authentication type. :type service_principal_id: any :param service_principal_key: The application's key used in AadServicePrincipal authentication type. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: The tenant information (domain name or tenant ID) used in AadServicePrincipal authentication type under which your application resides. :type tenant: any @@ -26892,9 +27099,9 @@ class RunFilterParameters(msrest.serialization.Model): 'ISO 8601' format. :type last_updated_before: ~datetime.datetime :param filters: List of filters. - :type filters: list[~azure.synapse.artifacts.v2021_06_01_preview.models.RunQueryFilter] + :type filters: list[~azure.synapse.artifacts.models.RunQueryFilter] :param order_by: List of OrderBy option. - :type order_by: list[~azure.synapse.artifacts.v2021_06_01_preview.models.RunQueryOrderBy] + :type order_by: list[~azure.synapse.artifacts.models.RunQueryOrderBy] """ _validation = { @@ -26933,11 +27140,10 @@ class RunQueryFilter(msrest.serialization.Model): runs are TriggerName, TriggerRunTimestamp and Status. Possible values include: "PipelineName", "Status", "RunStart", "RunEnd", "ActivityName", "ActivityRunStart", "ActivityRunEnd", "ActivityType", "TriggerName", "TriggerRunTimestamp", "RunGroupId", "LatestOnly". - :type operand: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.RunQueryFilterOperand + :type operand: str or ~azure.synapse.artifacts.models.RunQueryFilterOperand :param operator: Required. Operator to be used for filter. Possible values include: "Equals", "NotEquals", "In", "NotIn". - :type operator: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.RunQueryFilterOperator + :type operator: str or ~azure.synapse.artifacts.models.RunQueryFilterOperator :param values: Required. List of filter values. :type values: list[str] """ @@ -26975,9 +27181,9 @@ class RunQueryOrderBy(msrest.serialization.Model): TriggerRunTimestamp and Status. Possible values include: "RunStart", "RunEnd", "PipelineName", "Status", "ActivityName", "ActivityRunStart", "ActivityRunEnd", "TriggerName", "TriggerRunTimestamp". - :type order_by: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.RunQueryOrderByField + :type order_by: str or ~azure.synapse.artifacts.models.RunQueryOrderByField :param order: Required. Sorting order of the parameter. Possible values include: "ASC", "DESC". - :type order: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.RunQueryOrder + :type order: str or ~azure.synapse.artifacts.models.RunQueryOrder """ _validation = { @@ -27010,13 +27216,11 @@ class SalesforceLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param environment_url: The URL of Salesforce instance. Default is @@ -27028,9 +27232,9 @@ class SalesforceLinkedService(LinkedService): (or Expression with resultType string). :type username: any :param password: The password for Basic authentication of the Salesforce instance. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param security_token: The security token is optional to remotely access Salesforce instance. - :type security_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type security_token: ~azure.synapse.artifacts.models.SecretBase :param api_version: The Salesforce API version used in ADF. Type: string (or Expression with resultType string). :type api_version: any @@ -27084,13 +27288,11 @@ class SalesforceMarketingCloudLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_properties: Properties used to connect to Salesforce Marketing Cloud. It is @@ -27101,7 +27303,7 @@ class SalesforceMarketingCloudLinkedService(LinkedService): :type client_id: any :param client_secret: The client secret associated with the Salesforce Marketing Cloud application. Type: string (or Expression with resultType string). - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type client_secret: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. Type: boolean (or Expression with resultType boolean). :type use_encrypted_endpoints: any @@ -27174,16 +27376,14 @@ class SalesforceMarketingCloudObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -27288,16 +27488,14 @@ class SalesforceObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param object_api_name: The Salesforce object API name. Type: string (or Expression with resultType string). :type object_api_name: any @@ -27341,13 +27539,11 @@ class SalesforceServiceCloudLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param environment_url: The URL of Salesforce Service Cloud instance. Default is @@ -27359,9 +27555,9 @@ class SalesforceServiceCloudLinkedService(LinkedService): (or Expression with resultType string). :type username: any :param password: The password for Basic authentication of the Salesforce instance. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param security_token: The security token is optional to remotely access Salesforce instance. - :type security_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type security_token: ~azure.synapse.artifacts.models.SecretBase :param api_version: The Salesforce API version used in ADF. Type: string (or Expression with resultType string). :type api_version: any @@ -27428,16 +27624,14 @@ class SalesforceServiceCloudObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param object_api_name: The Salesforce Service Cloud object API name. Type: string (or Expression with resultType string). :type object_api_name: any @@ -27497,8 +27691,7 @@ class SalesforceServiceCloudSink(CopySink): :type max_concurrent_connections: any :param write_behavior: The write behavior for the operation. Default is Insert. Possible values include: "Insert", "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SalesforceSinkWriteBehavior + :type write_behavior: str or ~azure.synapse.artifacts.models.SalesforceSinkWriteBehavior :param external_id_field_name: The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). :type external_id_field_name: any @@ -27562,8 +27755,7 @@ class SalesforceServiceCloudSource(CopySource): :type query: any :param read_behavior: The read behavior for the operation. Default is Query. Possible values include: "Query", "QueryAll". - :type read_behavior: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SalesforceSourceReadBehavior + :type read_behavior: str or ~azure.synapse.artifacts.models.SalesforceSourceReadBehavior :param additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :type additional_columns: any @@ -27622,8 +27814,7 @@ class SalesforceSink(CopySink): :type max_concurrent_connections: any :param write_behavior: The write behavior for the operation. Default is Insert. Possible values include: "Insert", "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SalesforceSinkWriteBehavior + :type write_behavior: str or ~azure.synapse.artifacts.models.SalesforceSinkWriteBehavior :param external_id_field_name: The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). :type external_id_field_name: any @@ -27693,8 +27884,7 @@ class SalesforceSource(TabularSource): :type query: any :param read_behavior: The read behavior for the operation. Default is Query. Possible values include: "Query", "QueryAll". - :type read_behavior: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SalesforceSourceReadBehavior + :type read_behavior: str or ~azure.synapse.artifacts.models.SalesforceSourceReadBehavior """ _validation = { @@ -27742,16 +27932,14 @@ class SapBwCubeDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder """ _validation = { @@ -27790,13 +27978,11 @@ class SapBWLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param server: Required. Host name of the SAP BW instance. Type: string (or Expression with @@ -27812,7 +27998,7 @@ class SapBWLinkedService(LinkedService): resultType string). :type user_name: any :param password: Password to access the SAP BW server. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -27919,13 +28105,11 @@ class SapCloudForCustomerLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param url: Required. The URL of SAP Cloud for Customer OData API. For example, @@ -27936,7 +28120,7 @@ class SapCloudForCustomerLinkedService(LinkedService): resultType string). :type username: any :param password: The password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Either encryptedCredential or username/password must be provided. Type: string (or Expression with resultType string). @@ -27992,16 +28176,14 @@ class SapCloudForCustomerResourceDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param path: Required. The path of the SAP Cloud for Customer OData entity. Type: string (or Expression with resultType string). :type path: any @@ -28063,7 +28245,7 @@ class SapCloudForCustomerSink(CopySink): :param write_behavior: The write behavior for the operation. Default is 'Insert'. Possible values include: "Insert", "Update". :type write_behavior: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SapCloudForCustomerSinkWriteBehavior + ~azure.synapse.artifacts.models.SapCloudForCustomerSinkWriteBehavior :param http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: @@ -28169,13 +28351,11 @@ class SapEccLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param url: Required. The URL of SAP ECC OData API. For example, @@ -28186,7 +28366,7 @@ class SapEccLinkedService(LinkedService): resultType string). :type username: str :param password: The password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Either encryptedCredential or username/password must be provided. Type: string (or Expression with resultType string). @@ -28242,16 +28422,14 @@ class SapEccResourceDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param path: Required. The path of the SAP ECC OData entity. Type: string (or Expression with resultType string). :type path: any @@ -28357,13 +28535,11 @@ class SapHanaLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: SAP HANA ODBC connection string. Type: string, SecureString or @@ -28374,13 +28550,12 @@ class SapHanaLinkedService(LinkedService): :type server: any :param authentication_type: The authentication type to be used to connect to the SAP HANA server. Possible values include: "Basic", "Windows". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SapHanaAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.SapHanaAuthenticationType :param user_name: Username to access the SAP HANA server. Type: string (or Expression with resultType string). :type user_name: any :param password: Password to access the SAP HANA server. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -28473,12 +28648,10 @@ class SapHanaSource(TabularSource): :type packet_size: any :param partition_option: The partition mechanism that will be used for SAP HANA read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "SapHanaDynamicRange". - :type partition_option: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SapHanaPartitionOption + :type partition_option: str or ~azure.synapse.artifacts.models.SapHanaPartitionOption :param partition_settings: The settings that will be leveraged for SAP HANA source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SapHanaPartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.SapHanaPartitionSettings """ _validation = { @@ -28530,16 +28703,14 @@ class SapHanaTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param schema_type_properties_schema: The schema name of SAP HANA. Type: string (or Expression with resultType string). :type schema_type_properties_schema: any @@ -28587,13 +28758,11 @@ class SapOpenHubLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param server: Required. Host name of the SAP BW instance where the open hub destination is @@ -28618,7 +28787,7 @@ class SapOpenHubLinkedService(LinkedService): :type user_name: any :param password: Password to access the SAP BW server where the open hub destination is located. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param message_server: The hostname of the SAP Message Server. Type: string (or Expression with resultType string). :type message_server: any @@ -28770,16 +28939,14 @@ class SapOpenHubTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param open_hub_destination_name: Required. The name of the Open Hub Destination with destination type as Database Table. Type: string (or Expression with resultType string). :type open_hub_destination_name: any @@ -28835,13 +29002,11 @@ class SapTableLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param server: Host name of the SAP instance where the table is located. Type: string (or @@ -28865,7 +29030,7 @@ class SapTableLinkedService(LinkedService): (or Expression with resultType string). :type user_name: any :param password: Password to access the SAP server where the table is located. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param message_server: The hostname of the SAP Message Server. Type: string (or Expression with resultType string). :type message_server: any @@ -29005,16 +29170,14 @@ class SapTableResourceDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: Required. The name of the SAP Table. Type: string (or Expression with resultType string). :type table_name: any @@ -29098,12 +29261,10 @@ class SapTableSource(TabularSource): :param partition_option: The partition mechanism that will be used for SAP table read in parallel. Possible values include: "None", "PartitionOnInt", "PartitionOnCalendarYear", "PartitionOnCalendarMonth", "PartitionOnCalendarDate", "PartitionOnTime". - :type partition_option: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SapTablePartitionOption + :type partition_option: str or ~azure.synapse.artifacts.models.SapTablePartitionOption :param partition_settings: The settings that will be leveraged for SAP table source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SapTablePartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.SapTablePartitionSettings """ _validation = { @@ -29162,15 +29323,13 @@ class ScheduleTrigger(MultiplePipelineTrigger): :type description: str :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRuntimeState + :vartype runtime_state: str or ~azure.synapse.artifacts.models.TriggerRuntimeState :param annotations: List of tags that can be used for describing the trigger. :type annotations: list[any] :param pipelines: Pipelines that need to be started. - :type pipelines: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerPipelineReference] + :type pipelines: list[~azure.synapse.artifacts.models.TriggerPipelineReference] :param recurrence: Required. Recurrence schedule configuration. - :type recurrence: ~azure.synapse.artifacts.v2021_06_01_preview.models.ScheduleTriggerRecurrence + :type recurrence: ~azure.synapse.artifacts.models.ScheduleTriggerRecurrence """ _validation = { @@ -29206,7 +29365,7 @@ class ScheduleTriggerRecurrence(msrest.serialization.Model): :type additional_properties: dict[str, any] :param frequency: The frequency. Possible values include: "NotSpecified", "Minute", "Hour", "Day", "Week", "Month", "Year". - :type frequency: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.RecurrenceFrequency + :type frequency: str or ~azure.synapse.artifacts.models.RecurrenceFrequency :param interval: The interval. :type interval: int :param start_time: The start time. @@ -29216,7 +29375,7 @@ class ScheduleTriggerRecurrence(msrest.serialization.Model): :param time_zone: The time zone. :type time_zone: str :param schedule: The recurrence schedule. - :type schedule: ~azure.synapse.artifacts.v2021_06_01_preview.models.RecurrenceSchedule + :type schedule: ~azure.synapse.artifacts.models.RecurrenceSchedule """ _attribute_map = { @@ -29252,9 +29411,8 @@ class ScriptAction(msrest.serialization.Model): :type name: str :param uri: Required. The URI for the script action. :type uri: str - :param roles: Required. The node types on which the script action should be executed. Possible - values include: "Headnode", "Workernode", "Zookeeper". - :type roles: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.HdiNodeTypes + :param roles: Required. The node types on which the script action should be executed. + :type roles: any :param parameters: The parameters for the script action. :type parameters: str """ @@ -29268,7 +29426,7 @@ class ScriptAction(msrest.serialization.Model): _attribute_map = { 'name': {'key': 'name', 'type': 'str'}, 'uri': {'key': 'uri', 'type': 'str'}, - 'roles': {'key': 'roles', 'type': 'str'}, + 'roles': {'key': 'roles', 'type': 'object'}, 'parameters': {'key': 'parameters', 'type': 'str'}, } @@ -29360,12 +29518,11 @@ class SelfHostedIntegrationRuntime(IntegrationRuntime): :type additional_properties: dict[str, any] :param type: Required. Type of integration runtime.Constant filled by server. Possible values include: "Managed", "SelfHosted". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeType + :type type: str or ~azure.synapse.artifacts.models.IntegrationRuntimeType :param description: Integration runtime description. :type description: str :param linked_info: Linked integration runtime type from data factory. - :type linked_info: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedIntegrationRuntimeType + :type linked_info: ~azure.synapse.artifacts.models.LinkedIntegrationRuntimeType """ _validation = { @@ -29399,13 +29556,11 @@ class ServiceNowLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param endpoint: Required. The endpoint of the ServiceNow server. (i.e. @@ -29413,18 +29568,17 @@ class ServiceNowLinkedService(LinkedService): :type endpoint: any :param authentication_type: Required. The authentication type to use. Possible values include: "Basic", "OAuth2". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.ServiceNowAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.ServiceNowAuthenticationType :param username: The user name used to connect to the ServiceNow server for Basic and OAuth2 authentication. :type username: any :param password: The password corresponding to the user name for Basic and OAuth2 authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param client_id: The client id for OAuth2 authentication. :type client_id: any :param client_secret: The client secret for OAuth2 authentication. - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type client_secret: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -29503,16 +29657,14 @@ class ServiceNowObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -29613,9 +29765,9 @@ class SetVariableActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param variable_name: Name of the variable whose value needs to be set. :type variable_name: str :param value: Value to be set. Could be a static value or Expression. @@ -29774,13 +29926,11 @@ class SftpServerLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The SFTP server host name. Type: string (or Expression with resultType @@ -29791,13 +29941,12 @@ class SftpServerLinkedService(LinkedService): :type port: any :param authentication_type: The authentication type to be used to connect to the FTP server. Possible values include: "Basic", "SshPublicKey". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SftpAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.SftpAuthenticationType :param user_name: The username used to log on to the SFTP server. Type: string (or Expression with resultType string). :type user_name: any :param password: Password to logon the SFTP server for Basic authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -29810,10 +29959,10 @@ class SftpServerLinkedService(LinkedService): :param private_key_content: Base64 encoded SSH private key content for SshPublicKey authentication. For on-premises copy with SshPublicKey authentication, either PrivateKeyPath or PrivateKeyContent should be specified. SSH private key should be OpenSSH format. - :type private_key_content: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type private_key_content: ~azure.synapse.artifacts.models.SecretBase :param pass_phrase: The password to decrypt the SSH private key if the SSH private key is encrypted. - :type pass_phrase: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type pass_phrase: ~azure.synapse.artifacts.models.SecretBase :param skip_host_key_validation: If true, skip the SSH host key validation. Default value is false. Type: boolean (or Expression with resultType boolean). :type skip_host_key_validation: any @@ -29925,13 +30074,11 @@ class SharePointOnlineListLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param site_url: Required. The URL of the SharePoint Online site. For example, @@ -29948,7 +30095,7 @@ class SharePointOnlineListLinkedService(LinkedService): :type service_principal_id: any :param service_principal_key: Required. The client secret of your application registered in Azure Active Directory. Type: string (or Expression with resultType string). - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -30009,16 +30156,14 @@ class SharePointOnlineListResourceDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param list_name: The name of the SharePoint Online list. Type: string (or Expression with resultType string). :type list_name: any @@ -30114,20 +30259,18 @@ class ShopifyLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The endpoint of the Shopify server. (i.e. mystore.myshopify.com). :type host: any :param access_token: The API access token that can be used to access Shopify’s data. The token won't expire if it is offline mode. - :type access_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type access_token: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -30197,16 +30340,14 @@ class ShopifyObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -30364,16 +30505,14 @@ class SnowflakeDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param schema_type_properties_schema: The schema name of the Snowflake database. Type: string (or Expression with resultType string). :type schema_type_properties_schema: any @@ -30508,21 +30647,18 @@ class SnowflakeLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The connection string of snowflake. Type: string, SecureString. :type connection_string: any :param password: The Azure key vault secret reference of password in connection string. - :type password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type password: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -30586,8 +30722,7 @@ class SnowflakeSink(CopySink): string). :type pre_copy_script: any :param import_settings: Snowflake import settings. - :type import_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SnowflakeImportCopyCommand + :type import_settings: ~azure.synapse.artifacts.models.SnowflakeImportCopyCommand """ _validation = { @@ -30638,8 +30773,7 @@ class SnowflakeSource(CopySource): :param query: Snowflake Sql query. Type: string (or Expression with resultType string). :type query: any :param export_settings: Snowflake export settings. - :type export_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SnowflakeExportCopyCommand + :type export_settings: ~azure.synapse.artifacts.models.SnowflakeExportCopyCommand """ _validation = { @@ -30672,7 +30806,7 @@ class SparkBatchJob(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param livy_info: - :type livy_info: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkBatchJobState + :type livy_info: ~azure.synapse.artifacts.models.SparkBatchJobState :param name: The batch name. :type name: str :param workspace_name: The workspace name. @@ -30686,17 +30820,16 @@ class SparkBatchJob(msrest.serialization.Model): :param artifact_id: The artifact identifier. :type artifact_id: str :param job_type: The job type. Possible values include: "SparkBatch", "SparkSession". - :type job_type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobType + :type job_type: str or ~azure.synapse.artifacts.models.SparkJobType :param result: The Spark batch job result. Possible values include: "Uncertain", "Succeeded", "Failed", "Cancelled". - :type result: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkBatchJobResultType + :type result: str or ~azure.synapse.artifacts.models.SparkBatchJobResultType :param scheduler: The scheduler information. - :type scheduler: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkScheduler + :type scheduler: ~azure.synapse.artifacts.models.SparkScheduler :param plugin: The plugin information. - :type plugin: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkServicePlugin + :type plugin: ~azure.synapse.artifacts.models.SparkServicePlugin :param errors: The error information. - :type errors: list[~azure.synapse.artifacts.v2021_06_01_preview.models.SparkServiceError] + :type errors: list[~azure.synapse.artifacts.models.SparkServiceError] :param tags: A set of tags. The tags. :type tags: dict[str, str] :param id: Required. The session Id. @@ -30705,8 +30838,9 @@ class SparkBatchJob(msrest.serialization.Model): :type app_id: str :param app_info: The detailed application info. :type app_info: dict[str, str] - :param state: The batch state. - :type state: str + :param state: The batch state. Possible values include: "not_started", "starting", "idle", + "busy", "shutting_down", "error", "dead", "killed", "success", "running", "recovering". + :type state: str or ~azure.synapse.artifacts.models.LivyStates :param log_lines: The log lines. :type log_lines: list[str] """ @@ -30781,7 +30915,7 @@ class SparkBatchJobState(msrest.serialization.Model): :param current_state: the Spark job state. :type current_state: str :param job_creation_request: - :type job_creation_request: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkRequest + :type job_creation_request: ~azure.synapse.artifacts.models.SparkRequest """ _attribute_map = { @@ -30829,6 +30963,8 @@ class SparkConfiguration(msrest.serialization.Model): :type created_by: str :param created: The timestamp of resource creation. :type created: ~datetime.datetime + :param config_merge_rule: SparkConfiguration configMergeRule. + :type config_merge_rule: dict[str, str] """ _validation = { @@ -30842,6 +30978,7 @@ class SparkConfiguration(msrest.serialization.Model): 'notes': {'key': 'notes', 'type': 'str'}, 'created_by': {'key': 'createdBy', 'type': 'str'}, 'created': {'key': 'created', 'type': 'iso-8601'}, + 'config_merge_rule': {'key': 'configMergeRule', 'type': '{str}'}, } def __init__( @@ -30855,6 +30992,7 @@ def __init__( self.notes = kwargs.get('notes', None) self.created_by = kwargs.get('created_by', None) self.created = kwargs.get('created', None) + self.config_merge_rule = kwargs.get('config_merge_rule', None) class SparkConfigurationListResponse(msrest.serialization.Model): @@ -30863,8 +31001,7 @@ class SparkConfigurationListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of sparkconfigurations. - :type value: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.SparkConfigurationResource] + :type value: list[~azure.synapse.artifacts.models.SparkConfigurationResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -30905,7 +31042,7 @@ class SparkConfigurationResource(SubResource): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Properties of Spark Configuration. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkConfiguration + :type properties: ~azure.synapse.artifacts.models.SparkConfiguration """ _validation = { @@ -30943,14 +31080,16 @@ class SparkJobDefinition(msrest.serialization.Model): :param description: The description of the Spark job definition. :type description: str :param target_big_data_pool: Required. Big data pool reference. - :type target_big_data_pool: - ~azure.synapse.artifacts.v2021_06_01_preview.models.BigDataPoolReference + :type target_big_data_pool: ~azure.synapse.artifacts.models.BigDataPoolReference :param required_spark_version: The required Spark version of the application. :type required_spark_version: str :param language: The language of the Spark application. :type language: str :param job_properties: Required. The properties of the Spark job. - :type job_properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobProperties + :type job_properties: ~azure.synapse.artifacts.models.SparkJobProperties + :param folder: The folder that this Spark job definition is in. If not specified, this Spark + job definition will appear at the root level. + :type folder: ~azure.synapse.artifacts.models.SparkJobDefinitionFolder """ _validation = { @@ -30965,6 +31104,7 @@ class SparkJobDefinition(msrest.serialization.Model): 'required_spark_version': {'key': 'requiredSparkVersion', 'type': 'str'}, 'language': {'key': 'language', 'type': 'str'}, 'job_properties': {'key': 'jobProperties', 'type': 'SparkJobProperties'}, + 'folder': {'key': 'folder', 'type': 'SparkJobDefinitionFolder'}, } def __init__( @@ -30978,6 +31118,26 @@ def __init__( self.required_spark_version = kwargs.get('required_spark_version', None) self.language = kwargs.get('language', None) self.job_properties = kwargs['job_properties'] + self.folder = kwargs.get('folder', None) + + +class SparkJobDefinitionFolder(msrest.serialization.Model): + """The folder that this Spark job definition is in. If not specified, this Spark job definition will appear at the root level. + + :param name: The name of the folder that this Spark job definition is in. + :type name: str + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + } + + def __init__( + self, + **kwargs + ): + super(SparkJobDefinitionFolder, self).__init__(**kwargs) + self.name = kwargs.get('name', None) class SparkJobDefinitionResource(SubResource): @@ -30998,7 +31158,7 @@ class SparkJobDefinitionResource(SubResource): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Properties of spark job definition. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobDefinition + :type properties: ~azure.synapse.artifacts.models.SparkJobDefinition """ _validation = { @@ -31031,8 +31191,7 @@ class SparkJobDefinitionsListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of spark job definitions. - :type value: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobDefinitionResource] + :type value: list[~azure.synapse.artifacts.models.SparkJobDefinitionResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -31149,13 +31308,11 @@ class SparkLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. IP address or host name of the Spark server. @@ -31165,21 +31322,20 @@ class SparkLinkedService(LinkedService): :type port: any :param server_type: The type of Spark server. Possible values include: "SharkServer", "SharkServer2", "SparkThriftServer". - :type server_type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkServerType + :type server_type: str or ~azure.synapse.artifacts.models.SparkServerType :param thrift_transport_protocol: The transport protocol to use in the Thrift layer. Possible values include: "Binary", "SASL", "HTTP ". :type thrift_transport_protocol: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkThriftTransportProtocol + ~azure.synapse.artifacts.models.SparkThriftTransportProtocol :param authentication_type: Required. The authentication method used to access the Spark server. Possible values include: "Anonymous", "Username", "UsernameAndPassword", "WindowsAzureHDInsightService". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.SparkAuthenticationType :param username: The user name that you use to access Spark Server. :type username: any :param password: The password corresponding to the user name that you provided in the Username field. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param http_path: The partial URL corresponding to the Spark server. :type http_path: any :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The @@ -31275,16 +31431,14 @@ class SparkObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -31409,8 +31563,7 @@ class SparkScheduler(msrest.serialization.Model): :param cancellation_requested_at: :type cancellation_requested_at: ~datetime.datetime :param current_state: Possible values include: "Queued", "Scheduled", "Ended". - :type current_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SchedulerCurrentState + :type current_state: str or ~azure.synapse.artifacts.models.SchedulerCurrentState """ _attribute_map = { @@ -31441,7 +31594,7 @@ class SparkServiceError(msrest.serialization.Model): :param error_code: :type error_code: str :param source: Possible values include: "System", "User", "Unknown", "Dependency". - :type source: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkErrorSource + :type source: str or ~azure.synapse.artifacts.models.SparkErrorSource """ _attribute_map = { @@ -31475,8 +31628,7 @@ class SparkServicePlugin(msrest.serialization.Model): :type cleanup_started_at: ~datetime.datetime :param current_state: Possible values include: "Preparation", "ResourceAcquisition", "Queued", "Submission", "Monitoring", "Cleanup", "Ended". - :type current_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.PluginCurrentState + :type current_state: str or ~azure.synapse.artifacts.models.PluginCurrentState """ _attribute_map = { @@ -31558,27 +31710,25 @@ def __init__( class SqlConnection(msrest.serialization.Model): """The connection used to execute the SQL script. - All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. The type of the connection. Possible values include: "SqlOnDemand", - "SqlPool". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlConnectionType - :param name: Required. The identifier of the connection. + :param type: The type of the connection. Possible values include: "SqlOnDemand", "SqlPool". + :type type: str or ~azure.synapse.artifacts.models.SqlConnectionType + :param name: The identifier of the connection. :type name: str + :param pool_name: The associated SQL pool name (supported by SQL pool v3). + :type pool_name: str + :param database_name: The associated database name (supported by SQL pool v3). + :type database_name: str """ - _validation = { - 'type': {'required': True}, - 'name': {'required': True}, - } - _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, 'type': {'key': 'type', 'type': 'str'}, 'name': {'key': 'name', 'type': 'str'}, + 'pool_name': {'key': 'poolName', 'type': 'str'}, + 'database_name': {'key': 'databaseName', 'type': 'str'}, } def __init__( @@ -31587,8 +31737,10 @@ def __init__( ): super(SqlConnection, self).__init__(**kwargs) self.additional_properties = kwargs.get('additional_properties', None) - self.type = kwargs['type'] - self.name = kwargs['name'] + self.type = kwargs.get('type', None) + self.name = kwargs.get('name', None) + self.pool_name = kwargs.get('pool_name', None) + self.database_name = kwargs.get('database_name', None) class SqlDWSink(CopySink): @@ -31623,14 +31775,13 @@ class SqlDWSink(CopySink): applicable. Type: boolean (or Expression with resultType boolean). :type allow_poly_base: any :param poly_base_settings: Specifies PolyBase-related settings when allowPolyBase is true. - :type poly_base_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.PolybaseSettings + :type poly_base_settings: ~azure.synapse.artifacts.models.PolybaseSettings :param allow_copy_command: Indicates to use Copy Command to copy data into SQL Data Warehouse. Type: boolean (or Expression with resultType boolean). :type allow_copy_command: any :param copy_command_settings: Specifies Copy Command related settings when allowCopyCommand is true. - :type copy_command_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.DWCopyCommandSettings + :type copy_command_settings: ~azure.synapse.artifacts.models.DWCopyCommandSettings :param table_option: The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). :type table_option: any @@ -31710,8 +31861,7 @@ class SqlDWSource(TabularSource): Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". :type partition_option: any :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.SqlPartitionSettings """ _validation = { @@ -31782,7 +31932,7 @@ class SqlMISink(CopySink): :type pre_copy_script: any :param stored_procedure_parameters: SQL stored procedure parameters. :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] :param stored_procedure_table_type_parameter_name: The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). :type stored_procedure_table_type_parameter_name: any @@ -31859,15 +32009,14 @@ class SqlMISource(TabularSource): :param stored_procedure_parameters: Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] :param produce_additional_types: Which additional types to produce. :type produce_additional_types: any :param partition_option: The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". :type partition_option: any :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.SqlPartitionSettings """ _validation = { @@ -31960,7 +32109,7 @@ class SqlPool(TrackedResource): :param location: Required. The geo-location where the resource lives. :type location: str :param sku: SQL pool SKU. - :type sku: ~azure.synapse.artifacts.v2021_06_01_preview.models.Sku + :type sku: ~azure.synapse.artifacts.models.Sku :param max_size_bytes: Maximum size in bytes. :type max_size_bytes: long :param collation: Collation mode. @@ -32029,7 +32178,7 @@ class SqlPoolInfoListResult(msrest.serialization.Model): :param next_link: Link to the next page of results. :type next_link: str :param value: List of SQL pools. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPool] + :type value: list[~azure.synapse.artifacts.models.SqlPool] """ _attribute_map = { @@ -32052,7 +32201,7 @@ class SqlPoolReference(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param type: Required. SQL pool reference type. Possible values include: "SqlPoolReference". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPoolReferenceType + :type type: str or ~azure.synapse.artifacts.models.SqlPoolReferenceType :param reference_name: Required. Reference SQL pool name. :type reference_name: str """ @@ -32091,18 +32240,18 @@ class SqlPoolStoredProcedureActivity(Activity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param sql_pool: Required. SQL pool stored procedure reference. - :type sql_pool: ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPoolReference + :type sql_pool: ~azure.synapse.artifacts.models.SqlPoolReference :param stored_procedure_name: Required. Stored procedure name. Type: string (or Expression with resultType string). :type stored_procedure_name: any :param stored_procedure_parameters: Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] """ _validation = { @@ -32146,9 +32295,12 @@ class SqlScript(msrest.serialization.Model): :param description: The description of the SQL script. :type description: str :param type: The type of the SQL script. Possible values include: "SqlQuery". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlScriptType + :type type: str or ~azure.synapse.artifacts.models.SqlScriptType :param content: Required. The content of the SQL script. - :type content: ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlScriptContent + :type content: ~azure.synapse.artifacts.models.SqlScriptContent + :param folder: The folder that this SQL script is in. If not specified, this SQL script will + appear at the root level. + :type folder: ~azure.synapse.artifacts.models.SqlScriptFolder """ _validation = { @@ -32160,6 +32312,7 @@ class SqlScript(msrest.serialization.Model): 'description': {'key': 'description', 'type': 'str'}, 'type': {'key': 'type', 'type': 'str'}, 'content': {'key': 'content', 'type': 'SqlScriptContent'}, + 'folder': {'key': 'folder', 'type': 'SqlScriptFolder'}, } def __init__( @@ -32171,6 +32324,7 @@ def __init__( self.description = kwargs.get('description', None) self.type = kwargs.get('type', None) self.content = kwargs['content'] + self.folder = kwargs.get('folder', None) class SqlScriptContent(msrest.serialization.Model): @@ -32183,21 +32337,23 @@ class SqlScriptContent(msrest.serialization.Model): :type additional_properties: dict[str, any] :param query: Required. SQL query to execute. :type query: str - :param current_connection: Required. The connection used to execute the SQL script. - :type current_connection: ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlConnection + :param current_connection: The connection used to execute the SQL script. + :type current_connection: ~azure.synapse.artifacts.models.SqlConnection + :param result_limit: Limit of results, '-1' for no limit. + :type result_limit: int :param metadata: The metadata of the SQL script. - :type metadata: ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlScriptMetadata + :type metadata: ~azure.synapse.artifacts.models.SqlScriptMetadata """ _validation = { 'query': {'required': True}, - 'current_connection': {'required': True}, } _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, 'query': {'key': 'query', 'type': 'str'}, 'current_connection': {'key': 'currentConnection', 'type': 'SqlConnection'}, + 'result_limit': {'key': 'resultLimit', 'type': 'int'}, 'metadata': {'key': 'metadata', 'type': 'SqlScriptMetadata'}, } @@ -32208,10 +32364,30 @@ def __init__( super(SqlScriptContent, self).__init__(**kwargs) self.additional_properties = kwargs.get('additional_properties', None) self.query = kwargs['query'] - self.current_connection = kwargs['current_connection'] + self.current_connection = kwargs.get('current_connection', None) + self.result_limit = kwargs.get('result_limit', None) self.metadata = kwargs.get('metadata', None) +class SqlScriptFolder(msrest.serialization.Model): + """The folder that this SQL script is in. If not specified, this SQL script will appear at the root level. + + :param name: The name of the folder that this SQL script is in. + :type name: str + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + } + + def __init__( + self, + **kwargs + ): + super(SqlScriptFolder, self).__init__(**kwargs) + self.name = kwargs.get('name', None) + + class SqlScriptMetadata(msrest.serialization.Model): """The metadata of the SQL script. @@ -32254,7 +32430,7 @@ class SqlScriptResource(msrest.serialization.Model): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Properties of sql script. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlScript + :type properties: ~azure.synapse.artifacts.models.SqlScript """ _validation = { @@ -32291,7 +32467,7 @@ class SqlScriptsListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of sql scripts. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.SqlScriptResource] + :type value: list[~azure.synapse.artifacts.models.SqlScriptResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -32325,13 +32501,11 @@ class SqlServerLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The connection string. Type: string, SecureString or @@ -32341,7 +32515,7 @@ class SqlServerLinkedService(LinkedService): with resultType string). :type user_name: any :param password: The on-premises Windows authentication password. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -32414,7 +32588,7 @@ class SqlServerSink(CopySink): :type pre_copy_script: any :param stored_procedure_parameters: SQL stored procedure parameters. :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] :param stored_procedure_table_type_parameter_name: The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). :type stored_procedure_table_type_parameter_name: any @@ -32491,15 +32665,14 @@ class SqlServerSource(TabularSource): :param stored_procedure_parameters: Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] :param produce_additional_types: Which additional types to produce. :type produce_additional_types: any :param partition_option: The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". :type partition_option: any :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.SqlPartitionSettings """ _validation = { @@ -32551,21 +32724,20 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param stored_procedure_name: Required. Stored procedure name. Type: string (or Expression with resultType string). :type stored_procedure_name: any :param stored_procedure_parameters: Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] """ _validation = { @@ -32616,16 +32788,14 @@ class SqlServerTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -32704,7 +32874,7 @@ class SqlSink(CopySink): :type pre_copy_script: any :param stored_procedure_parameters: SQL stored procedure parameters. :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] :param stored_procedure_table_type_parameter_name: The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). :type stored_procedure_table_type_parameter_name: any @@ -32781,7 +32951,7 @@ class SqlSource(TabularSource): :param stored_procedure_parameters: Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] :param isolation_level: Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). @@ -32790,8 +32960,7 @@ class SqlSource(TabularSource): Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". :type partition_option: any :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.SqlPartitionSettings """ _validation = { @@ -32839,13 +33008,11 @@ class SquareLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_properties: Properties used to connect to Square. It is mutually exclusive @@ -32856,7 +33023,7 @@ class SquareLinkedService(LinkedService): :param client_id: Required. The client ID associated with your Square application. :type client_id: any :param client_secret: The client secret associated with your Square application. - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type client_secret: ~azure.synapse.artifacts.models.SecretBase :param redirect_uri: Required. The redirect URL assigned in the Square application dashboard. (i.e. http://localhost:2500). :type redirect_uri: any @@ -32937,16 +33104,14 @@ class SquareObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -33042,7 +33207,7 @@ class SSISAccessCredential(msrest.serialization.Model): :param user_name: Required. UseName for windows authentication. :type user_name: any :param password: Required. Password for windows authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase """ _validation = { @@ -33117,7 +33282,7 @@ class SSISExecutionCredential(msrest.serialization.Model): :param user_name: Required. UseName for windows authentication. :type user_name: any :param password: Required. Password for windows authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecureString + :type password: ~azure.synapse.artifacts.models.SecureString """ _validation = { @@ -33177,10 +33342,9 @@ class SSISLogLocation(msrest.serialization.Model): with resultType string). :type log_path: any :param type: Required. The type of SSIS log location. Possible values include: "File". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.SsisLogLocationType + :type type: str or ~azure.synapse.artifacts.models.SsisLogLocationType :param access_credential: The package execution log access credential. - :type access_credential: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISAccessCredential + :type access_credential: ~azure.synapse.artifacts.models.SSISAccessCredential :param log_refresh_interval: Specifies the interval to refresh log. The default interval is 5 minutes. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). @@ -33249,18 +33413,16 @@ class SSISPackageLocation(msrest.serialization.Model): :type package_path: any :param type: The type of SSIS package location. Possible values include: "SSISDB", "File", "InlinePackage", "PackageStore". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.SsisPackageLocationType + :type type: str or ~azure.synapse.artifacts.models.SsisPackageLocationType :param package_password: Password of the package. - :type package_password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type package_password: ~azure.synapse.artifacts.models.SecretBase :param access_credential: The package access credential. - :type access_credential: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISAccessCredential + :type access_credential: ~azure.synapse.artifacts.models.SSISAccessCredential :param configuration_path: The configuration file of the package execution. Type: string (or Expression with resultType string). :type configuration_path: any :param configuration_access_credential: The configuration file access credential. - :type configuration_access_credential: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISAccessCredential + :type configuration_access_credential: ~azure.synapse.artifacts.models.SSISAccessCredential :param package_name: The package name. :type package_name: str :param package_content: The embedded package content. Type: string (or Expression with @@ -33269,8 +33431,7 @@ class SSISPackageLocation(msrest.serialization.Model): :param package_last_modified_date: The embedded package last modified date. :type package_last_modified_date: str :param child_packages: The embedded child package list. - :type child_packages: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.SSISChildPackage] + :type child_packages: list[~azure.synapse.artifacts.models.SSISChildPackage] """ _attribute_map = { @@ -33343,8 +33504,7 @@ class StagingSettings(msrest.serialization.Model): collection. :type additional_properties: dict[str, any] :param linked_service_name: Required. Staging linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param path: The path to storage for storing the interim data. Type: string (or Expression with resultType string). :type path: any @@ -33381,12 +33541,11 @@ class StartDataFlowDebugSessionRequest(msrest.serialization.Model): :param session_id: The ID of data flow debug session. :type session_id: str :param data_flow: Data flow instance. - :type data_flow: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowResource + :type data_flow: ~azure.synapse.artifacts.models.DataFlowResource :param datasets: List of datasets. - :type datasets: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetResource] + :type datasets: list[~azure.synapse.artifacts.models.DatasetResource] :param linked_services: List of linked services. - :type linked_services: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceResource] + :type linked_services: list[~azure.synapse.artifacts.models.LinkedServiceResource] :param staging: Staging info for debug session. :type staging: any :param debug_settings: Data flow debug settings. @@ -33446,8 +33605,7 @@ class StoredProcedureParameter(msrest.serialization.Model): :type value: any :param type: Stored procedure parameter type. Possible values include: "String", "Int", "Int64", "Decimal", "Guid", "Boolean", "Date". - :type type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameterType + :type type: str or ~azure.synapse.artifacts.models.StoredProcedureParameterType """ _attribute_map = { @@ -33479,19 +33637,19 @@ class SwitchActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param on: Required. An expression that would evaluate to a string or integer. This is used to determine the block of activities in cases that will be executed. - :type on: ~azure.synapse.artifacts.v2021_06_01_preview.models.Expression + :type on: ~azure.synapse.artifacts.models.Expression :param cases: List of cases that correspond to expected values of the 'on' property. This is an optional property and if not provided, the activity will execute activities provided in defaultActivities. - :type cases: list[~azure.synapse.artifacts.v2021_06_01_preview.models.SwitchCase] + :type cases: list[~azure.synapse.artifacts.models.SwitchCase] :param default_activities: List of activities to execute if no case condition is satisfied. This is an optional property and if not provided, the activity will exit without any action. - :type default_activities: list[~azure.synapse.artifacts.v2021_06_01_preview.models.Activity] + :type default_activities: list[~azure.synapse.artifacts.models.Activity] """ _validation = { @@ -33529,7 +33687,7 @@ class SwitchCase(msrest.serialization.Model): :param value: Expected value that satisfies the expression result of the 'on' property. :type value: str :param activities: List of activities to execute for satisfied case condition. - :type activities: list[~azure.synapse.artifacts.v2021_06_01_preview.models.Activity] + :type activities: list[~azure.synapse.artifacts.models.Activity] """ _attribute_map = { @@ -33557,13 +33715,11 @@ class SybaseLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param server: Required. Server name for connection. Type: string (or Expression with @@ -33576,13 +33732,12 @@ class SybaseLinkedService(LinkedService): :type schema: any :param authentication_type: AuthenticationType to be used for connection. Possible values include: "Basic", "Windows". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SybaseAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.SybaseAuthenticationType :param username: Username for authentication. Type: string (or Expression with resultType string). :type username: any :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -33698,16 +33853,14 @@ class SybaseTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The Sybase table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -33754,16 +33907,15 @@ class SynapseNotebookActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param notebook: Required. Synapse notebook reference. - :type notebook: ~azure.synapse.artifacts.v2021_06_01_preview.models.SynapseNotebookReference + :type notebook: ~azure.synapse.artifacts.models.SynapseNotebookReference :param parameters: Notebook parameters. :type parameters: dict[str, any] """ @@ -33804,7 +33956,7 @@ class SynapseNotebookReference(msrest.serialization.Model): :param type: Required. Synapse notebook reference type. Possible values include: "NotebookReference". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookReferenceType + :type type: str or ~azure.synapse.artifacts.models.NotebookReferenceType :param reference_name: Required. Reference notebook name. :type reference_name: str """ @@ -33843,16 +33995,15 @@ class SynapseSparkJobDefinitionActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param spark_job: Required. Synapse spark job reference. - :type spark_job: ~azure.synapse.artifacts.v2021_06_01_preview.models.SynapseSparkJobReference + :type spark_job: ~azure.synapse.artifacts.models.SynapseSparkJobReference :param arguments: User specified arguments to SynapseSparkJobDefinitionActivity. :type arguments: list[any] """ @@ -33893,7 +34044,7 @@ class SynapseSparkJobReference(msrest.serialization.Model): :param type: Required. Synapse spark job reference type. Possible values include: "SparkJobDefinitionReference". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobReferenceType + :type type: str or ~azure.synapse.artifacts.models.SparkJobReferenceType :param reference_name: Required. Reference spark job name. :type reference_name: str """ @@ -33952,8 +34103,7 @@ class TabularTranslator(CopyTranslator): activity. Type: boolean (or Expression with resultType boolean). :type type_conversion: any :param type_conversion_settings: Type conversion settings. - :type type_conversion_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.TypeConversionSettings + :type type_conversion_settings: ~azure.synapse.artifacts.models.TypeConversionSettings """ _validation = { @@ -34066,13 +34216,11 @@ class TeradataLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Teradata ODBC connection string. Type: string, SecureString or @@ -34082,13 +34230,12 @@ class TeradataLinkedService(LinkedService): :type server: any :param authentication_type: AuthenticationType to be used for connection. Possible values include: "Basic", "Windows". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TeradataAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.TeradataAuthenticationType :param username: Username for authentication. Type: string (or Expression with resultType string). :type username: any :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -34189,12 +34336,10 @@ class TeradataSource(TabularSource): :type query: any :param partition_option: The partition mechanism that will be used for teradata read in parallel. Possible values include: "None", "Hash", "DynamicRange". - :type partition_option: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TeradataPartitionOption + :type partition_option: str or ~azure.synapse.artifacts.models.TeradataPartitionOption :param partition_settings: The settings that will be leveraged for teradata source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.TeradataPartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.TeradataPartitionSettings """ _validation = { @@ -34244,16 +34389,14 @@ class TeradataTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param database: The database name of Teradata. Type: string (or Expression with resultType string). :type database: any @@ -34411,7 +34554,7 @@ class TriggerDependencyReference(DependencyReference): :param type: Required. The type of dependency reference.Constant filled by server. :type type: str :param reference_trigger: Required. Referenced trigger. - :type reference_trigger: ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerReference + :type reference_trigger: ~azure.synapse.artifacts.models.TriggerReference """ _validation = { @@ -34443,7 +34586,7 @@ class TriggerListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of triggers. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerResource] + :type value: list[~azure.synapse.artifacts.models.TriggerResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -34470,7 +34613,7 @@ class TriggerPipelineReference(msrest.serialization.Model): """Pipeline that needs to be triggered with the given parameters. :param pipeline_reference: Pipeline reference. - :type pipeline_reference: ~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineReference + :type pipeline_reference: ~azure.synapse.artifacts.models.PipelineReference :param parameters: Pipeline parameters. :type parameters: dict[str, any] """ @@ -34495,7 +34638,7 @@ class TriggerReference(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param type: Required. Trigger reference type. Possible values include: "TriggerReference". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerReferenceType + :type type: str or ~azure.synapse.artifacts.models.TriggerReferenceType :param reference_name: Required. Reference trigger name. :type reference_name: str """ @@ -34537,7 +34680,7 @@ class TriggerResource(SubResource): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Properties of the trigger. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.Trigger + :type properties: ~azure.synapse.artifacts.models.Trigger """ _validation = { @@ -34581,7 +34724,7 @@ class TriggerRun(msrest.serialization.Model): :ivar trigger_run_timestamp: Trigger run start time. :vartype trigger_run_timestamp: ~datetime.datetime :ivar status: Trigger run status. Possible values include: "Succeeded", "Failed", "Inprogress". - :vartype status: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRunStatus + :vartype status: str or ~azure.synapse.artifacts.models.TriggerRunStatus :ivar message: Trigger error message. :vartype message: str :ivar properties: List of property name and value related to trigger run. Name, value pair @@ -34636,7 +34779,7 @@ class TriggerRunsQueryResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of trigger runs. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRun] + :type value: list[~azure.synapse.artifacts.models.TriggerRun] :param continuation_token: The continuation token for getting the next page of results, if any remaining results exist, null otherwise. :type continuation_token: str @@ -34669,8 +34812,7 @@ class TriggerSubscriptionOperationStatus(msrest.serialization.Model): :vartype trigger_name: str :ivar status: Event Subscription Status. Possible values include: "Enabled", "Provisioning", "Deprovisioning", "Disabled", "Unknown". - :vartype status: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.EventSubscriptionStatus + :vartype status: str or ~azure.synapse.artifacts.models.EventSubscriptionStatus """ _validation = { @@ -34708,17 +34850,15 @@ class TumblingWindowTrigger(Trigger): :type description: str :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRuntimeState + :vartype runtime_state: str or ~azure.synapse.artifacts.models.TriggerRuntimeState :param annotations: List of tags that can be used for describing the trigger. :type annotations: list[any] :param pipeline: Required. Pipeline for which runs are created when an event is fired for trigger window that is ready. - :type pipeline: ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerPipelineReference + :type pipeline: ~azure.synapse.artifacts.models.TriggerPipelineReference :param frequency: Required. The frequency of the time windows. Possible values include: "Minute", "Hour", "Month". - :type frequency: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TumblingWindowFrequency + :type frequency: str or ~azure.synapse.artifacts.models.TumblingWindowFrequency :param interval: Required. The interval of the time windows. The minimum interval allowed is 15 Minutes. :type interval: int @@ -34736,10 +34876,10 @@ class TumblingWindowTrigger(Trigger): for which a new run is triggered. :type max_concurrency: int :param retry_policy: Retry policy that will be applied for failed pipeline runs. - :type retry_policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.RetryPolicy + :type retry_policy: ~azure.synapse.artifacts.models.RetryPolicy :param depends_on: Triggers that this trigger depends on. Only tumbling window triggers are supported. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DependencyReference] + :type depends_on: list[~azure.synapse.artifacts.models.DependencyReference] """ _validation = { @@ -34794,7 +34934,7 @@ class TumblingWindowTriggerDependencyReference(TriggerDependencyReference): :param type: Required. The type of dependency reference.Constant filled by server. :type type: str :param reference_trigger: Required. Referenced trigger. - :type reference_trigger: ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerReference + :type reference_trigger: ~azure.synapse.artifacts.models.TriggerReference :param offset: Timespan applied to the start time of a tumbling window when evaluating dependency. :type offset: str @@ -34887,12 +35027,12 @@ class UntilActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param expression: Required. An expression that would evaluate to Boolean. The loop will continue until this expression evaluates to true. - :type expression: ~azure.synapse.artifacts.v2021_06_01_preview.models.Expression + :type expression: ~azure.synapse.artifacts.models.Expression :param timeout: Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or Expression with resultType string), pattern: @@ -34900,7 +35040,7 @@ class UntilActivity(ControlActivity): resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type timeout: any :param activities: Required. List of activities to execute. - :type activities: list[~azure.synapse.artifacts.v2021_06_01_preview.models.Activity] + :type activities: list[~azure.synapse.artifacts.models.Activity] """ _validation = { @@ -34979,9 +35119,9 @@ class ValidationActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param timeout: Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or Expression with resultType string), pattern: @@ -34998,7 +35138,7 @@ class ValidationActivity(ControlActivity): with resultType boolean). :type child_items: any :param dataset: Required. Validation activity dataset reference. - :type dataset: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference + :type dataset: ~azure.synapse.artifacts.models.DatasetReference """ _validation = { @@ -35041,7 +35181,7 @@ class VariableSpecification(msrest.serialization.Model): :param type: Required. Variable type. Possible values include: "String", "Bool", "Boolean", "Array". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.VariableType + :type type: str or ~azure.synapse.artifacts.models.VariableType :param default_value: Default value of variable. :type default_value: any """ @@ -35075,20 +35215,18 @@ class VerticaLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type pwd: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -35195,16 +35333,14 @@ class VerticaTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -35281,9 +35417,9 @@ class WaitActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param wait_time_in_seconds: Required. Duration in seconds. :type wait_time_in_seconds: any """ @@ -35328,17 +35464,16 @@ class WebActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param method: Required. Rest API method for target endpoint. Possible values include: "GET", "POST", "PUT", "DELETE". - :type method: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.WebActivityMethod + :type method: str or ~azure.synapse.artifacts.models.WebActivityMethod :param url: Required. Web activity target endpoint and path. Type: string (or Expression with resultType string). :type url: any @@ -35350,16 +35485,13 @@ class WebActivity(ExecutionActivity): method, not allowed for GET method Type: string (or Expression with resultType string). :type body: any :param authentication: Authentication method used for calling the endpoint. - :type authentication: - ~azure.synapse.artifacts.v2021_06_01_preview.models.WebActivityAuthentication + :type authentication: ~azure.synapse.artifacts.models.WebActivityAuthentication :param datasets: List of datasets passed to web endpoint. - :type datasets: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference] + :type datasets: list[~azure.synapse.artifacts.models.DatasetReference] :param linked_services: List of linked services passed to web endpoint. - :type linked_services: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference] + :type linked_services: list[~azure.synapse.artifacts.models.LinkedServiceReference] :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference """ _validation = { @@ -35412,11 +35544,11 @@ class WebActivityAuthentication(msrest.serialization.Model): :param type: Required. Web activity authentication (Basic/ClientCertificate/MSI). :type type: str :param pfx: Base64-encoded contents of a PFX file. - :type pfx: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type pfx: ~azure.synapse.artifacts.models.SecretBase :param username: Web activity authentication user name for basic authentication. :type username: str :param password: Password for the PFX file or basic authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param resource: Resource for which Azure Auth token will be requested when using MSI Authentication. :type resource: str @@ -35460,8 +35592,7 @@ class WebLinkedServiceTypeProperties(msrest.serialization.Model): :param authentication_type: Required. Type of authentication used to connect to the web table source.Constant filled by server. Possible values include: "Basic", "Anonymous", "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.WebAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.WebAuthenticationType """ _validation = { @@ -35498,8 +35629,7 @@ class WebAnonymousAuthentication(WebLinkedServiceTypeProperties): :param authentication_type: Required. Type of authentication used to connect to the web table source.Constant filled by server. Possible values include: "Basic", "Anonymous", "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.WebAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.WebAuthenticationType """ _validation = { @@ -35531,13 +35661,12 @@ class WebBasicAuthentication(WebLinkedServiceTypeProperties): :param authentication_type: Required. Type of authentication used to connect to the web table source.Constant filled by server. Possible values include: "Basic", "Anonymous", "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.WebAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.WebAuthenticationType :param username: Required. User name for Basic authentication. Type: string (or Expression with resultType string). :type username: any :param password: Required. The password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase """ _validation = { @@ -35575,12 +35704,11 @@ class WebClientCertificateAuthentication(WebLinkedServiceTypeProperties): :param authentication_type: Required. Type of authentication used to connect to the web table source.Constant filled by server. Possible values include: "Basic", "Anonymous", "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.WebAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.WebAuthenticationType :param pfx: Required. Base64-encoded contents of a PFX file. - :type pfx: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type pfx: ~azure.synapse.artifacts.models.SecretBase :param password: Required. Password for the PFX file. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase """ _validation = { @@ -35622,11 +35750,11 @@ class WebHookActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param method: Required. Rest API method for target endpoint. Possible values include: "POST". - :type method: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.WebHookActivityMethod + :type method: str or ~azure.synapse.artifacts.models.WebHookActivityMethod :param url: Required. WebHook activity target endpoint and path. Type: string (or Expression with resultType string). :type url: any @@ -35642,8 +35770,7 @@ class WebHookActivity(ControlActivity): method, not allowed for GET method Type: string (or Expression with resultType string). :type body: any :param authentication: Authentication method used for calling the endpoint. - :type authentication: - ~azure.synapse.artifacts.v2021_06_01_preview.models.WebActivityAuthentication + :type authentication: ~azure.synapse.artifacts.models.WebActivityAuthentication :param report_status_on_call_back: When set to true, statusCode, output and error in callback request body will be consumed by activity. The activity can be marked as failed by setting statusCode >= 400 in callback request. Default is false. Type: boolean (or Expression with @@ -35700,18 +35827,15 @@ class WebLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param type_properties: Required. Web linked service properties. - :type type_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.WebLinkedServiceTypeProperties + :type type_properties: ~azure.synapse.artifacts.models.WebLinkedServiceTypeProperties """ _validation = { @@ -35803,16 +35927,14 @@ class WebTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param index: Required. The zero-based index of the table in the web page. Type: integer (or Expression with resultType integer), minimum: 0. :type index: any @@ -35871,10 +35993,9 @@ class Workspace(TrackedResource): :param location: Required. The geo-location where the resource lives. :type location: str :param identity: Identity of the workspace. - :type identity: ~azure.synapse.artifacts.v2021_06_01_preview.models.ManagedIdentity + :type identity: ~azure.synapse.artifacts.models.ManagedIdentity :param default_data_lake_storage: Workspace default data lake storage account details. - :type default_data_lake_storage: - ~azure.synapse.artifacts.v2021_06_01_preview.models.DataLakeStorageAccountDetails + :type default_data_lake_storage: ~azure.synapse.artifacts.models.DataLakeStorageAccountDetails :param sql_administrator_login_password: SQL administrator login password. :type sql_administrator_login_password: str :param managed_resource_group_name: Workspace managed resource group. The resource group name @@ -35887,8 +36008,7 @@ class Workspace(TrackedResource): :param sql_administrator_login: Login for workspace SQL active directory administrator. :type sql_administrator_login: str :param virtual_network_profile: Virtual Network profile. - :type virtual_network_profile: - ~azure.synapse.artifacts.v2021_06_01_preview.models.VirtualNetworkProfile + :type virtual_network_profile: ~azure.synapse.artifacts.models.VirtualNetworkProfile :param connectivity_endpoints: Connectivity endpoints. :type connectivity_endpoints: dict[str, str] :param managed_virtual_network: Setting this to 'default' will ensure that all compute for this @@ -35896,22 +36016,21 @@ class Workspace(TrackedResource): :type managed_virtual_network: str :param private_endpoint_connections: Private endpoint connections to the workspace. :type private_endpoint_connections: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.PrivateEndpointConnection] + list[~azure.synapse.artifacts.models.PrivateEndpointConnection] :param encryption: The encryption details of the workspace. - :type encryption: ~azure.synapse.artifacts.v2021_06_01_preview.models.EncryptionDetails + :type encryption: ~azure.synapse.artifacts.models.EncryptionDetails :ivar workspace_uid: The workspace unique identifier. :vartype workspace_uid: str :ivar extra_properties: Workspace level configs and feature flags. :vartype extra_properties: dict[str, any] :param managed_virtual_network_settings: Managed Virtual Network Settings. :type managed_virtual_network_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.ManagedVirtualNetworkSettings + ~azure.synapse.artifacts.models.ManagedVirtualNetworkSettings :param workspace_repository_configuration: Git integration settings. :type workspace_repository_configuration: - ~azure.synapse.artifacts.v2021_06_01_preview.models.WorkspaceRepositoryConfiguration + ~azure.synapse.artifacts.models.WorkspaceRepositoryConfiguration :param purview_configuration: Purview Configuration. - :type purview_configuration: - ~azure.synapse.artifacts.v2021_06_01_preview.models.PurviewConfiguration + :type purview_configuration: ~azure.synapse.artifacts.models.PurviewConfiguration :ivar adla_resource_id: The ADLA resource ID. :vartype adla_resource_id: str """ @@ -36063,7 +36182,7 @@ class WorkspaceRepositoryConfiguration(msrest.serialization.Model): :param client_id: GitHub bring your own app client id. :type client_id: str :param client_secret: GitHub bring your own app client secret information. - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.GitHubClientSecret + :type client_secret: ~azure.synapse.artifacts.models.GitHubClientSecret """ _attribute_map = { @@ -36104,7 +36223,7 @@ class WorkspaceUpdateParameters(msrest.serialization.Model): :param tags: A set of tags. The resource tags. :type tags: dict[str, str] :param identity: Managed service identity of the workspace. - :type identity: ~azure.synapse.artifacts.v2021_06_01_preview.models.WorkspaceIdentity + :type identity: ~azure.synapse.artifacts.models.WorkspaceIdentity """ _attribute_map = { @@ -36132,13 +36251,11 @@ class XeroLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_properties: Properties used to connect to Xero. It is mutually exclusive with @@ -36147,11 +36264,11 @@ class XeroLinkedService(LinkedService): :param host: Required. The endpoint of the Xero server. (i.e. api.xero.com). :type host: any :param consumer_key: The consumer key associated with the Xero application. - :type consumer_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type consumer_key: ~azure.synapse.artifacts.models.SecretBase :param private_key: The private key from the .pem file that was generated for your Xero private application. You must include all the text from the .pem file, including the Unix line endings( ). - :type private_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type private_key: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -36225,16 +36342,14 @@ class XeroObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -36339,18 +36454,16 @@ class XmlDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param location: The location of the json data storage. - :type location: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetLocation + :type location: ~azure.synapse.artifacts.models.DatasetLocation :param encoding_name: The code page name of the preferred encoding. If not specified, the default value is UTF-8, unless BOM denotes another Unicode encoding. Refer to the name column of the table in the following link to set supported values: @@ -36360,7 +36473,7 @@ class XmlDataset(Dataset): :param null_value: The null value string. Type: string (or Expression with resultType string). :type null_value: any :param compression: The data compression method used for the json dataset. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression """ _validation = { @@ -36407,8 +36520,7 @@ class XmlReadSettings(FormatReadSettings): :param type: Required. The read setting type.Constant filled by server. :type type: str :param compression_properties: Compression settings. - :type compression_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.CompressionReadSettings + :type compression_properties: ~azure.synapse.artifacts.models.CompressionReadSettings :param validation_mode: Indicates what validation method is used when reading the xml files. Allowed values: 'none', 'xsd', or 'dtd'. Type: string (or Expression with resultType string). :type validation_mode: any @@ -36472,9 +36584,9 @@ class XmlSource(CopySource): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Xml store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings :param format_settings: Xml format settings. - :type format_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.XmlReadSettings + :type format_settings: ~azure.synapse.artifacts.models.XmlReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :type additional_columns: any @@ -36551,13 +36663,11 @@ class ZohoLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_properties: Properties used to connect to Zoho. It is mutually exclusive with @@ -36566,7 +36676,7 @@ class ZohoLinkedService(LinkedService): :param endpoint: Required. The endpoint of the Zoho server. (i.e. crm.zoho.com/crm/private). :type endpoint: any :param access_token: The access token for Zoho authentication. - :type access_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type access_token: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -36638,16 +36748,14 @@ class ZohoObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/models/_models_py3.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/models/_models_py3.py similarity index 92% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/models/_models_py3.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/models/_models_py3.py index 667fcbd569e0..043c5920ef4d 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/models/_models_py3.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/models/_models_py3.py @@ -33,9 +33,9 @@ class Activity(msrest.serialization.Model): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] """ _validation = { @@ -86,8 +86,7 @@ class ActivityDependency(msrest.serialization.Model): :param activity: Required. Activity name. :type activity: str :param dependency_conditions: Required. Match-Condition for the dependency. - :type dependency_conditions: list[str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DependencyCondition] + :type dependency_conditions: list[str or ~azure.synapse.artifacts.models.DependencyCondition] """ _validation = { @@ -270,7 +269,7 @@ class ActivityRunsQueryResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of activity runs. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityRun] + :type value: list[~azure.synapse.artifacts.models.ActivityRun] :param continuation_token: The continuation token for getting the next page of results, if any remaining results exist, null otherwise. :type continuation_token: str @@ -348,7 +347,7 @@ class LinkedService(msrest.serialization.Model): """The Azure Synapse nested object which contains the information and credential which can be used to connect with related store or compute resource. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonMWSLinkedService, AmazonRedshiftLinkedService, AmazonS3LinkedService, AzureBatchLinkedService, AzureBlobFSLinkedService, AzureBlobStorageLinkedService, AzureDataExplorerLinkedService, AzureDataLakeAnalyticsLinkedService, AzureDataLakeStoreLinkedService, AzureDatabricksLinkedService, AzureDatabricksDeltaLakeLinkedService, AzureFileStorageLinkedService, AzureFunctionLinkedService, AzureKeyVaultLinkedService, AzureMLLinkedService, AzureMLServiceLinkedService, AzureMariaDBLinkedService, AzureMySqlLinkedService, AzurePostgreSqlLinkedService, AzureSearchLinkedService, AzureSqlDWLinkedService, AzureSqlDatabaseLinkedService, AzureSqlMILinkedService, AzureStorageLinkedService, AzureTableStorageLinkedService, CassandraLinkedService, CommonDataServiceForAppsLinkedService, ConcurLinkedService, CosmosDbLinkedService, CosmosDbMongoDbApiLinkedService, CouchbaseLinkedService, CustomDataSourceLinkedService, Db2LinkedService, DrillLinkedService, DynamicsLinkedService, DynamicsAXLinkedService, DynamicsCrmLinkedService, EloquaLinkedService, FileServerLinkedService, FtpServerLinkedService, GoogleAdWordsLinkedService, GoogleBigQueryLinkedService, GoogleCloudStorageLinkedService, GreenplumLinkedService, HBaseLinkedService, HDInsightLinkedService, HDInsightOnDemandLinkedService, HdfsLinkedService, HiveLinkedService, HttpLinkedService, HubspotLinkedService, ImpalaLinkedService, InformixLinkedService, JiraLinkedService, MagentoLinkedService, MariaDBLinkedService, MarketoLinkedService, MicrosoftAccessLinkedService, MongoDbLinkedService, MongoDbAtlasLinkedService, MongoDbV2LinkedService, MySqlLinkedService, NetezzaLinkedService, ODataLinkedService, OdbcLinkedService, Office365LinkedService, OracleLinkedService, OracleServiceCloudLinkedService, PaypalLinkedService, PhoenixLinkedService, PostgreSqlLinkedService, PrestoLinkedService, QuickBooksLinkedService, ResponsysLinkedService, RestServiceLinkedService, SalesforceLinkedService, SalesforceMarketingCloudLinkedService, SalesforceServiceCloudLinkedService, SapBWLinkedService, SapCloudForCustomerLinkedService, SapEccLinkedService, SapHanaLinkedService, SapOpenHubLinkedService, SapTableLinkedService, ServiceNowLinkedService, SftpServerLinkedService, SharePointOnlineListLinkedService, ShopifyLinkedService, SnowflakeLinkedService, SparkLinkedService, SqlServerLinkedService, SquareLinkedService, SybaseLinkedService, TeradataLinkedService, VerticaLinkedService, WebLinkedService, XeroLinkedService, ZohoLinkedService. + sub-classes are: AmazonMWSLinkedService, AmazonRdsForOracleLinkedService, AmazonRdsForSqlServerLinkedService, AmazonRedshiftLinkedService, AmazonS3LinkedService, AzureBatchLinkedService, AzureBlobFSLinkedService, AzureBlobStorageLinkedService, AzureDataExplorerLinkedService, AzureDataLakeAnalyticsLinkedService, AzureDataLakeStoreLinkedService, AzureDatabricksLinkedService, AzureDatabricksDeltaLakeLinkedService, AzureFileStorageLinkedService, AzureFunctionLinkedService, AzureKeyVaultLinkedService, AzureMLLinkedService, AzureMLServiceLinkedService, AzureMariaDBLinkedService, AzureMySqlLinkedService, AzurePostgreSqlLinkedService, AzureSearchLinkedService, AzureSqlDWLinkedService, AzureSqlDatabaseLinkedService, AzureSqlMILinkedService, AzureStorageLinkedService, AzureTableStorageLinkedService, CassandraLinkedService, CommonDataServiceForAppsLinkedService, ConcurLinkedService, CosmosDbLinkedService, CosmosDbMongoDbApiLinkedService, CouchbaseLinkedService, CustomDataSourceLinkedService, Db2LinkedService, DrillLinkedService, DynamicsLinkedService, DynamicsAXLinkedService, DynamicsCrmLinkedService, EloquaLinkedService, FileServerLinkedService, FtpServerLinkedService, GoogleAdWordsLinkedService, GoogleBigQueryLinkedService, GoogleCloudStorageLinkedService, GreenplumLinkedService, HBaseLinkedService, HDInsightLinkedService, HDInsightOnDemandLinkedService, HdfsLinkedService, HiveLinkedService, HttpLinkedService, HubspotLinkedService, ImpalaLinkedService, InformixLinkedService, JiraLinkedService, MagentoLinkedService, MariaDBLinkedService, MarketoLinkedService, MicrosoftAccessLinkedService, MongoDbLinkedService, MongoDbAtlasLinkedService, MongoDbV2LinkedService, MySqlLinkedService, NetezzaLinkedService, ODataLinkedService, OdbcLinkedService, Office365LinkedService, OracleLinkedService, OracleServiceCloudLinkedService, PaypalLinkedService, PhoenixLinkedService, PostgreSqlLinkedService, PrestoLinkedService, QuickBooksLinkedService, ResponsysLinkedService, RestServiceLinkedService, SalesforceLinkedService, SalesforceMarketingCloudLinkedService, SalesforceServiceCloudLinkedService, SapBWLinkedService, SapCloudForCustomerLinkedService, SapEccLinkedService, SapHanaLinkedService, SapOpenHubLinkedService, SapTableLinkedService, ServiceNowLinkedService, SftpServerLinkedService, SharePointOnlineListLinkedService, ShopifyLinkedService, SnowflakeLinkedService, SparkLinkedService, SqlServerLinkedService, SquareLinkedService, SybaseLinkedService, TeradataLinkedService, VerticaLinkedService, WebLinkedService, XeroLinkedService, ZohoLinkedService. All required parameters must be populated in order to send to Azure. @@ -358,13 +357,11 @@ class LinkedService(msrest.serialization.Model): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] """ @@ -383,7 +380,7 @@ class LinkedService(msrest.serialization.Model): } _subtype_map = { - 'type': {'AmazonMWS': 'AmazonMWSLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureBlobFS': 'AzureBlobFSLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureDataExplorer': 'AzureDataExplorerLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDatabricksDeltaLake': 'AzureDatabricksDeltaLakeLinkedService', 'AzureFileStorage': 'AzureFileStorageLinkedService', 'AzureFunction': 'AzureFunctionLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureML': 'AzureMLLinkedService', 'AzureMLService': 'AzureMLServiceLinkedService', 'AzureMariaDB': 'AzureMariaDBLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'AzureSqlMI': 'AzureSqlMILinkedService', 'AzureStorage': 'AzureStorageLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'Cassandra': 'CassandraLinkedService', 'CommonDataServiceForApps': 'CommonDataServiceForAppsLinkedService', 'Concur': 'ConcurLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'CosmosDbMongoDbApi': 'CosmosDbMongoDbApiLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'Db2': 'Db2LinkedService', 'Drill': 'DrillLinkedService', 'Dynamics': 'DynamicsLinkedService', 'DynamicsAX': 'DynamicsAXLinkedService', 'DynamicsCrm': 'DynamicsCrmLinkedService', 'Eloqua': 'EloquaLinkedService', 'FileServer': 'FileServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'GoogleAdWords': 'GoogleAdWordsLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'GoogleCloudStorage': 'GoogleCloudStorageLinkedService', 'Greenplum': 'GreenplumLinkedService', 'HBase': 'HBaseLinkedService', 'HDInsight': 'HDInsightLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'Hdfs': 'HdfsLinkedService', 'Hive': 'HiveLinkedService', 'HttpServer': 'HttpLinkedService', 'Hubspot': 'HubspotLinkedService', 'Impala': 'ImpalaLinkedService', 'Informix': 'InformixLinkedService', 'Jira': 'JiraLinkedService', 'Magento': 'MagentoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Marketo': 'MarketoLinkedService', 'MicrosoftAccess': 'MicrosoftAccessLinkedService', 'MongoDb': 'MongoDbLinkedService', 'MongoDbAtlas': 'MongoDbAtlasLinkedService', 'MongoDbV2': 'MongoDbV2LinkedService', 'MySql': 'MySqlLinkedService', 'Netezza': 'NetezzaLinkedService', 'OData': 'ODataLinkedService', 'Odbc': 'OdbcLinkedService', 'Office365': 'Office365LinkedService', 'Oracle': 'OracleLinkedService', 'OracleServiceCloud': 'OracleServiceCloudLinkedService', 'Paypal': 'PaypalLinkedService', 'Phoenix': 'PhoenixLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'Presto': 'PrestoLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Responsys': 'ResponsysLinkedService', 'RestService': 'RestServiceLinkedService', 'Salesforce': 'SalesforceLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'SalesforceServiceCloud': 'SalesforceServiceCloudLinkedService', 'SapBW': 'SapBWLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'SapEcc': 'SapEccLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapOpenHub': 'SapOpenHubLinkedService', 'SapTable': 'SapTableLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'Sftp': 'SftpServerLinkedService', 'SharePointOnlineList': 'SharePointOnlineListLinkedService', 'Shopify': 'ShopifyLinkedService', 'Snowflake': 'SnowflakeLinkedService', 'Spark': 'SparkLinkedService', 'SqlServer': 'SqlServerLinkedService', 'Square': 'SquareLinkedService', 'Sybase': 'SybaseLinkedService', 'Teradata': 'TeradataLinkedService', 'Vertica': 'VerticaLinkedService', 'Web': 'WebLinkedService', 'Xero': 'XeroLinkedService', 'Zoho': 'ZohoLinkedService'} + 'type': {'AmazonMWS': 'AmazonMWSLinkedService', 'AmazonRdsForOracle': 'AmazonRdsForOracleLinkedService', 'AmazonRdsForSqlServer': 'AmazonRdsForSqlServerLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureBlobFS': 'AzureBlobFSLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureDataExplorer': 'AzureDataExplorerLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDatabricksDeltaLake': 'AzureDatabricksDeltaLakeLinkedService', 'AzureFileStorage': 'AzureFileStorageLinkedService', 'AzureFunction': 'AzureFunctionLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureML': 'AzureMLLinkedService', 'AzureMLService': 'AzureMLServiceLinkedService', 'AzureMariaDB': 'AzureMariaDBLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'AzureSqlMI': 'AzureSqlMILinkedService', 'AzureStorage': 'AzureStorageLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'Cassandra': 'CassandraLinkedService', 'CommonDataServiceForApps': 'CommonDataServiceForAppsLinkedService', 'Concur': 'ConcurLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'CosmosDbMongoDbApi': 'CosmosDbMongoDbApiLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'Db2': 'Db2LinkedService', 'Drill': 'DrillLinkedService', 'Dynamics': 'DynamicsLinkedService', 'DynamicsAX': 'DynamicsAXLinkedService', 'DynamicsCrm': 'DynamicsCrmLinkedService', 'Eloqua': 'EloquaLinkedService', 'FileServer': 'FileServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'GoogleAdWords': 'GoogleAdWordsLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'GoogleCloudStorage': 'GoogleCloudStorageLinkedService', 'Greenplum': 'GreenplumLinkedService', 'HBase': 'HBaseLinkedService', 'HDInsight': 'HDInsightLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'Hdfs': 'HdfsLinkedService', 'Hive': 'HiveLinkedService', 'HttpServer': 'HttpLinkedService', 'Hubspot': 'HubspotLinkedService', 'Impala': 'ImpalaLinkedService', 'Informix': 'InformixLinkedService', 'Jira': 'JiraLinkedService', 'Magento': 'MagentoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Marketo': 'MarketoLinkedService', 'MicrosoftAccess': 'MicrosoftAccessLinkedService', 'MongoDb': 'MongoDbLinkedService', 'MongoDbAtlas': 'MongoDbAtlasLinkedService', 'MongoDbV2': 'MongoDbV2LinkedService', 'MySql': 'MySqlLinkedService', 'Netezza': 'NetezzaLinkedService', 'OData': 'ODataLinkedService', 'Odbc': 'OdbcLinkedService', 'Office365': 'Office365LinkedService', 'Oracle': 'OracleLinkedService', 'OracleServiceCloud': 'OracleServiceCloudLinkedService', 'Paypal': 'PaypalLinkedService', 'Phoenix': 'PhoenixLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'Presto': 'PrestoLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Responsys': 'ResponsysLinkedService', 'RestService': 'RestServiceLinkedService', 'Salesforce': 'SalesforceLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'SalesforceServiceCloud': 'SalesforceServiceCloudLinkedService', 'SapBW': 'SapBWLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'SapEcc': 'SapEccLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapOpenHub': 'SapOpenHubLinkedService', 'SapTable': 'SapTableLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'Sftp': 'SftpServerLinkedService', 'SharePointOnlineList': 'SharePointOnlineListLinkedService', 'Shopify': 'ShopifyLinkedService', 'Snowflake': 'SnowflakeLinkedService', 'Spark': 'SparkLinkedService', 'SqlServer': 'SqlServerLinkedService', 'Square': 'SquareLinkedService', 'Sybase': 'SybaseLinkedService', 'Teradata': 'TeradataLinkedService', 'Vertica': 'VerticaLinkedService', 'Web': 'WebLinkedService', 'Xero': 'XeroLinkedService', 'Zoho': 'ZohoLinkedService'} } def __init__( @@ -416,13 +413,11 @@ class AmazonMWSLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param endpoint: Required. The endpoint of the Amazon MWS server, (i.e. @@ -435,11 +430,11 @@ class AmazonMWSLinkedService(LinkedService): :param seller_id: Required. The Amazon seller ID. :type seller_id: any :param mws_auth_token: The Amazon MWS authentication token. - :type mws_auth_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type mws_auth_token: ~azure.synapse.artifacts.models.SecretBase :param access_key_id: Required. The access key id used to access data. :type access_key_id: any :param secret_key: The secret key used to access data. - :type secret_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type secret_key: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -521,7 +516,7 @@ class Dataset(msrest.serialization.Model): """The Azure Data Factory nested object which identifies data within different data stores, such as tables, files, folders, and documents. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonMWSObjectDataset, AmazonRedshiftTableDataset, AmazonS3Dataset, AvroDataset, AzureBlobDataset, AzureBlobFSDataset, AzureDataExplorerTableDataset, AzureDataLakeStoreDataset, AzureDatabricksDeltaLakeDataset, AzureMariaDBTableDataset, AzureMySqlTableDataset, AzurePostgreSqlTableDataset, AzureSearchIndexDataset, AzureSqlDWTableDataset, AzureSqlMITableDataset, AzureSqlTableDataset, AzureTableDataset, BinaryDataset, CassandraTableDataset, CommonDataServiceForAppsEntityDataset, ConcurObjectDataset, CosmosDbMongoDbApiCollectionDataset, CosmosDbSqlApiCollectionDataset, CouchbaseTableDataset, CustomDataset, Db2TableDataset, DelimitedTextDataset, DocumentDbCollectionDataset, DrillTableDataset, DynamicsAXResourceDataset, DynamicsCrmEntityDataset, DynamicsEntityDataset, EloquaObjectDataset, ExcelDataset, FileShareDataset, GoogleAdWordsObjectDataset, GoogleBigQueryObjectDataset, GreenplumTableDataset, HBaseObjectDataset, HiveObjectDataset, HttpDataset, HubspotObjectDataset, ImpalaObjectDataset, InformixTableDataset, JiraObjectDataset, JsonDataset, MagentoObjectDataset, MariaDBTableDataset, MarketoObjectDataset, MicrosoftAccessTableDataset, MongoDbAtlasCollectionDataset, MongoDbCollectionDataset, MongoDbV2CollectionDataset, MySqlTableDataset, NetezzaTableDataset, ODataResourceDataset, OdbcTableDataset, Office365Dataset, OracleServiceCloudObjectDataset, OracleTableDataset, OrcDataset, ParquetDataset, PaypalObjectDataset, PhoenixObjectDataset, PostgreSqlTableDataset, PrestoObjectDataset, QuickBooksObjectDataset, RelationalTableDataset, ResponsysObjectDataset, RestResourceDataset, SalesforceMarketingCloudObjectDataset, SalesforceObjectDataset, SalesforceServiceCloudObjectDataset, SapBwCubeDataset, SapCloudForCustomerResourceDataset, SapEccResourceDataset, SapHanaTableDataset, SapOpenHubTableDataset, SapTableResourceDataset, ServiceNowObjectDataset, SharePointOnlineListResourceDataset, ShopifyObjectDataset, SnowflakeDataset, SparkObjectDataset, SqlServerTableDataset, SquareObjectDataset, SybaseTableDataset, TeradataTableDataset, VerticaTableDataset, WebTableDataset, XeroObjectDataset, XmlDataset, ZohoObjectDataset. + sub-classes are: AmazonMWSObjectDataset, AmazonRdsForOracleTableDataset, AmazonRdsForSqlServerTableDataset, AmazonRedshiftTableDataset, AmazonS3Dataset, AvroDataset, AzureBlobDataset, AzureBlobFSDataset, AzureDataExplorerTableDataset, AzureDataLakeStoreDataset, AzureDatabricksDeltaLakeDataset, AzureMariaDBTableDataset, AzureMySqlTableDataset, AzurePostgreSqlTableDataset, AzureSearchIndexDataset, AzureSqlDWTableDataset, AzureSqlMITableDataset, AzureSqlTableDataset, AzureTableDataset, BinaryDataset, CassandraTableDataset, CommonDataServiceForAppsEntityDataset, ConcurObjectDataset, CosmosDbMongoDbApiCollectionDataset, CosmosDbSqlApiCollectionDataset, CouchbaseTableDataset, CustomDataset, Db2TableDataset, DelimitedTextDataset, DocumentDbCollectionDataset, DrillTableDataset, DynamicsAXResourceDataset, DynamicsCrmEntityDataset, DynamicsEntityDataset, EloquaObjectDataset, ExcelDataset, FileShareDataset, GoogleAdWordsObjectDataset, GoogleBigQueryObjectDataset, GreenplumTableDataset, HBaseObjectDataset, HiveObjectDataset, HttpDataset, HubspotObjectDataset, ImpalaObjectDataset, InformixTableDataset, JiraObjectDataset, JsonDataset, MagentoObjectDataset, MariaDBTableDataset, MarketoObjectDataset, MicrosoftAccessTableDataset, MongoDbAtlasCollectionDataset, MongoDbCollectionDataset, MongoDbV2CollectionDataset, MySqlTableDataset, NetezzaTableDataset, ODataResourceDataset, OdbcTableDataset, Office365Dataset, OracleServiceCloudObjectDataset, OracleTableDataset, OrcDataset, ParquetDataset, PaypalObjectDataset, PhoenixObjectDataset, PostgreSqlTableDataset, PrestoObjectDataset, QuickBooksObjectDataset, RelationalTableDataset, ResponsysObjectDataset, RestResourceDataset, SalesforceMarketingCloudObjectDataset, SalesforceObjectDataset, SalesforceServiceCloudObjectDataset, SapBwCubeDataset, SapCloudForCustomerResourceDataset, SapEccResourceDataset, SapHanaTableDataset, SapOpenHubTableDataset, SapTableResourceDataset, ServiceNowObjectDataset, SharePointOnlineListResourceDataset, ShopifyObjectDataset, SnowflakeDataset, SparkObjectDataset, SqlServerTableDataset, SquareObjectDataset, SybaseTableDataset, TeradataTableDataset, VerticaTableDataset, WebTableDataset, XeroObjectDataset, XmlDataset, ZohoObjectDataset. All required parameters must be populated in order to send to Azure. @@ -539,16 +534,14 @@ class Dataset(msrest.serialization.Model): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder """ _validation = { @@ -569,7 +562,7 @@ class Dataset(msrest.serialization.Model): } _subtype_map = { - 'type': {'AmazonMWSObject': 'AmazonMWSObjectDataset', 'AmazonRedshiftTable': 'AmazonRedshiftTableDataset', 'AmazonS3Object': 'AmazonS3Dataset', 'Avro': 'AvroDataset', 'AzureBlob': 'AzureBlobDataset', 'AzureBlobFSFile': 'AzureBlobFSDataset', 'AzureDataExplorerTable': 'AzureDataExplorerTableDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'AzureDatabricksDeltaLakeDataset': 'AzureDatabricksDeltaLakeDataset', 'AzureMariaDBTable': 'AzureMariaDBTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlMITable': 'AzureSqlMITableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'Binary': 'BinaryDataset', 'CassandraTable': 'CassandraTableDataset', 'CommonDataServiceForAppsEntity': 'CommonDataServiceForAppsEntityDataset', 'ConcurObject': 'ConcurObjectDataset', 'CosmosDbMongoDbApiCollection': 'CosmosDbMongoDbApiCollectionDataset', 'CosmosDbSqlApiCollection': 'CosmosDbSqlApiCollectionDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'CustomDataset': 'CustomDataset', 'Db2Table': 'Db2TableDataset', 'DelimitedText': 'DelimitedTextDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'DrillTable': 'DrillTableDataset', 'DynamicsAXResource': 'DynamicsAXResourceDataset', 'DynamicsCrmEntity': 'DynamicsCrmEntityDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'EloquaObject': 'EloquaObjectDataset', 'Excel': 'ExcelDataset', 'FileShare': 'FileShareDataset', 'GoogleAdWordsObject': 'GoogleAdWordsObjectDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'HBaseObject': 'HBaseObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HttpFile': 'HttpDataset', 'HubspotObject': 'HubspotObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'InformixTable': 'InformixTableDataset', 'JiraObject': 'JiraObjectDataset', 'Json': 'JsonDataset', 'MagentoObject': 'MagentoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MarketoObject': 'MarketoObjectDataset', 'MicrosoftAccessTable': 'MicrosoftAccessTableDataset', 'MongoDbAtlasCollection': 'MongoDbAtlasCollectionDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'MongoDbV2Collection': 'MongoDbV2CollectionDataset', 'MySqlTable': 'MySqlTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ODataResource': 'ODataResourceDataset', 'OdbcTable': 'OdbcTableDataset', 'Office365Table': 'Office365Dataset', 'OracleServiceCloudObject': 'OracleServiceCloudObjectDataset', 'OracleTable': 'OracleTableDataset', 'Orc': 'OrcDataset', 'Parquet': 'ParquetDataset', 'PaypalObject': 'PaypalObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PostgreSqlTable': 'PostgreSqlTableDataset', 'PrestoObject': 'PrestoObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'ResponsysObject': 'ResponsysObjectDataset', 'RestResource': 'RestResourceDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'SalesforceServiceCloudObject': 'SalesforceServiceCloudObjectDataset', 'SapBwCube': 'SapBwCubeDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapHanaTable': 'SapHanaTableDataset', 'SapOpenHubTable': 'SapOpenHubTableDataset', 'SapTableResource': 'SapTableResourceDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'SharePointOnlineListResource': 'SharePointOnlineListResourceDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'SnowflakeTable': 'SnowflakeDataset', 'SparkObject': 'SparkObjectDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SquareObject': 'SquareObjectDataset', 'SybaseTable': 'SybaseTableDataset', 'TeradataTable': 'TeradataTableDataset', 'VerticaTable': 'VerticaTableDataset', 'WebTable': 'WebTableDataset', 'XeroObject': 'XeroObjectDataset', 'Xml': 'XmlDataset', 'ZohoObject': 'ZohoObjectDataset'} + 'type': {'AmazonMWSObject': 'AmazonMWSObjectDataset', 'AmazonRdsForOracleTable': 'AmazonRdsForOracleTableDataset', 'AmazonRdsForSqlServerTable': 'AmazonRdsForSqlServerTableDataset', 'AmazonRedshiftTable': 'AmazonRedshiftTableDataset', 'AmazonS3Object': 'AmazonS3Dataset', 'Avro': 'AvroDataset', 'AzureBlob': 'AzureBlobDataset', 'AzureBlobFSFile': 'AzureBlobFSDataset', 'AzureDataExplorerTable': 'AzureDataExplorerTableDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'AzureDatabricksDeltaLakeDataset': 'AzureDatabricksDeltaLakeDataset', 'AzureMariaDBTable': 'AzureMariaDBTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlMITable': 'AzureSqlMITableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'Binary': 'BinaryDataset', 'CassandraTable': 'CassandraTableDataset', 'CommonDataServiceForAppsEntity': 'CommonDataServiceForAppsEntityDataset', 'ConcurObject': 'ConcurObjectDataset', 'CosmosDbMongoDbApiCollection': 'CosmosDbMongoDbApiCollectionDataset', 'CosmosDbSqlApiCollection': 'CosmosDbSqlApiCollectionDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'CustomDataset': 'CustomDataset', 'Db2Table': 'Db2TableDataset', 'DelimitedText': 'DelimitedTextDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'DrillTable': 'DrillTableDataset', 'DynamicsAXResource': 'DynamicsAXResourceDataset', 'DynamicsCrmEntity': 'DynamicsCrmEntityDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'EloquaObject': 'EloquaObjectDataset', 'Excel': 'ExcelDataset', 'FileShare': 'FileShareDataset', 'GoogleAdWordsObject': 'GoogleAdWordsObjectDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'HBaseObject': 'HBaseObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HttpFile': 'HttpDataset', 'HubspotObject': 'HubspotObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'InformixTable': 'InformixTableDataset', 'JiraObject': 'JiraObjectDataset', 'Json': 'JsonDataset', 'MagentoObject': 'MagentoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MarketoObject': 'MarketoObjectDataset', 'MicrosoftAccessTable': 'MicrosoftAccessTableDataset', 'MongoDbAtlasCollection': 'MongoDbAtlasCollectionDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'MongoDbV2Collection': 'MongoDbV2CollectionDataset', 'MySqlTable': 'MySqlTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ODataResource': 'ODataResourceDataset', 'OdbcTable': 'OdbcTableDataset', 'Office365Table': 'Office365Dataset', 'OracleServiceCloudObject': 'OracleServiceCloudObjectDataset', 'OracleTable': 'OracleTableDataset', 'Orc': 'OrcDataset', 'Parquet': 'ParquetDataset', 'PaypalObject': 'PaypalObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PostgreSqlTable': 'PostgreSqlTableDataset', 'PrestoObject': 'PrestoObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'ResponsysObject': 'ResponsysObjectDataset', 'RestResource': 'RestResourceDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'SalesforceServiceCloudObject': 'SalesforceServiceCloudObjectDataset', 'SapBwCube': 'SapBwCubeDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapHanaTable': 'SapHanaTableDataset', 'SapOpenHubTable': 'SapOpenHubTableDataset', 'SapTableResource': 'SapTableResourceDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'SharePointOnlineListResource': 'SharePointOnlineListResourceDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'SnowflakeTable': 'SnowflakeDataset', 'SparkObject': 'SparkObjectDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SquareObject': 'SquareObjectDataset', 'SybaseTable': 'SybaseTableDataset', 'TeradataTable': 'TeradataTableDataset', 'VerticaTable': 'VerticaTableDataset', 'WebTable': 'WebTableDataset', 'XeroObject': 'XeroObjectDataset', 'Xml': 'XmlDataset', 'ZohoObject': 'ZohoObjectDataset'} } def __init__( @@ -616,16 +609,14 @@ class AmazonMWSObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -671,7 +662,7 @@ class CopySource(msrest.serialization.Model): """A copy activity source. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AvroSource, AzureBlobFSSource, AzureDataExplorerSource, AzureDataLakeStoreSource, AzureDatabricksDeltaLakeSource, BinarySource, BlobSource, CommonDataServiceForAppsSource, CosmosDbMongoDbApiSource, CosmosDbSqlApiSource, DelimitedTextSource, DocumentDbCollectionSource, DynamicsCrmSource, DynamicsSource, ExcelSource, FileSystemSource, HdfsSource, HttpSource, JsonSource, MicrosoftAccessSource, MongoDbAtlasSource, MongoDbSource, MongoDbV2Source, ODataSource, Office365Source, OracleSource, OrcSource, ParquetSource, RelationalSource, RestSource, SalesforceServiceCloudSource, SharePointOnlineListSource, SnowflakeSource, TabularSource, WebSource, XmlSource. + sub-classes are: AmazonRdsForOracleSource, AvroSource, AzureBlobFSSource, AzureDataExplorerSource, AzureDataLakeStoreSource, AzureDatabricksDeltaLakeSource, BinarySource, BlobSource, CommonDataServiceForAppsSource, CosmosDbMongoDbApiSource, CosmosDbSqlApiSource, DelimitedTextSource, DocumentDbCollectionSource, DynamicsCrmSource, DynamicsSource, ExcelSource, FileSystemSource, HdfsSource, HttpSource, JsonSource, MicrosoftAccessSource, MongoDbAtlasSource, MongoDbSource, MongoDbV2Source, ODataSource, Office365Source, OracleSource, OrcSource, ParquetSource, RelationalSource, RestSource, SalesforceServiceCloudSource, SharePointOnlineListSource, SnowflakeSource, TabularSource, WebSource, XmlSource. All required parameters must be populated in order to send to Azure. @@ -704,7 +695,7 @@ class CopySource(msrest.serialization.Model): } _subtype_map = { - 'type': {'AvroSource': 'AvroSource', 'AzureBlobFSSource': 'AzureBlobFSSource', 'AzureDataExplorerSource': 'AzureDataExplorerSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'AzureDatabricksDeltaLakeSource': 'AzureDatabricksDeltaLakeSource', 'BinarySource': 'BinarySource', 'BlobSource': 'BlobSource', 'CommonDataServiceForAppsSource': 'CommonDataServiceForAppsSource', 'CosmosDbMongoDbApiSource': 'CosmosDbMongoDbApiSource', 'CosmosDbSqlApiSource': 'CosmosDbSqlApiSource', 'DelimitedTextSource': 'DelimitedTextSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'DynamicsCrmSource': 'DynamicsCrmSource', 'DynamicsSource': 'DynamicsSource', 'ExcelSource': 'ExcelSource', 'FileSystemSource': 'FileSystemSource', 'HdfsSource': 'HdfsSource', 'HttpSource': 'HttpSource', 'JsonSource': 'JsonSource', 'MicrosoftAccessSource': 'MicrosoftAccessSource', 'MongoDbAtlasSource': 'MongoDbAtlasSource', 'MongoDbSource': 'MongoDbSource', 'MongoDbV2Source': 'MongoDbV2Source', 'ODataSource': 'ODataSource', 'Office365Source': 'Office365Source', 'OracleSource': 'OracleSource', 'OrcSource': 'OrcSource', 'ParquetSource': 'ParquetSource', 'RelationalSource': 'RelationalSource', 'RestSource': 'RestSource', 'SalesforceServiceCloudSource': 'SalesforceServiceCloudSource', 'SharePointOnlineListSource': 'SharePointOnlineListSource', 'SnowflakeSource': 'SnowflakeSource', 'TabularSource': 'TabularSource', 'WebSource': 'WebSource', 'XmlSource': 'XmlSource'} + 'type': {'AmazonRdsForOracleSource': 'AmazonRdsForOracleSource', 'AvroSource': 'AvroSource', 'AzureBlobFSSource': 'AzureBlobFSSource', 'AzureDataExplorerSource': 'AzureDataExplorerSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'AzureDatabricksDeltaLakeSource': 'AzureDatabricksDeltaLakeSource', 'BinarySource': 'BinarySource', 'BlobSource': 'BlobSource', 'CommonDataServiceForAppsSource': 'CommonDataServiceForAppsSource', 'CosmosDbMongoDbApiSource': 'CosmosDbMongoDbApiSource', 'CosmosDbSqlApiSource': 'CosmosDbSqlApiSource', 'DelimitedTextSource': 'DelimitedTextSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'DynamicsCrmSource': 'DynamicsCrmSource', 'DynamicsSource': 'DynamicsSource', 'ExcelSource': 'ExcelSource', 'FileSystemSource': 'FileSystemSource', 'HdfsSource': 'HdfsSource', 'HttpSource': 'HttpSource', 'JsonSource': 'JsonSource', 'MicrosoftAccessSource': 'MicrosoftAccessSource', 'MongoDbAtlasSource': 'MongoDbAtlasSource', 'MongoDbSource': 'MongoDbSource', 'MongoDbV2Source': 'MongoDbV2Source', 'ODataSource': 'ODataSource', 'Office365Source': 'Office365Source', 'OracleSource': 'OracleSource', 'OrcSource': 'OrcSource', 'ParquetSource': 'ParquetSource', 'RelationalSource': 'RelationalSource', 'RestSource': 'RestSource', 'SalesforceServiceCloudSource': 'SalesforceServiceCloudSource', 'SharePointOnlineListSource': 'SharePointOnlineListSource', 'SnowflakeSource': 'SnowflakeSource', 'TabularSource': 'TabularSource', 'WebSource': 'WebSource', 'XmlSource': 'XmlSource'} } def __init__( @@ -728,7 +719,7 @@ class TabularSource(CopySource): """Copy activity sources of tabular type. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonMWSSource, AmazonRedshiftSource, AzureMariaDBSource, AzureMySqlSource, AzurePostgreSqlSource, AzureSqlSource, AzureTableSource, CassandraSource, ConcurSource, CouchbaseSource, Db2Source, DrillSource, DynamicsAXSource, EloquaSource, GoogleAdWordsSource, GoogleBigQuerySource, GreenplumSource, HBaseSource, HiveSource, HubspotSource, ImpalaSource, InformixSource, JiraSource, MagentoSource, MariaDBSource, MarketoSource, MySqlSource, NetezzaSource, OdbcSource, OracleServiceCloudSource, PaypalSource, PhoenixSource, PostgreSqlSource, PrestoSource, QuickBooksSource, ResponsysSource, SalesforceMarketingCloudSource, SalesforceSource, SapBwSource, SapCloudForCustomerSource, SapEccSource, SapHanaSource, SapOpenHubSource, SapTableSource, ServiceNowSource, ShopifySource, SparkSource, SqlDWSource, SqlMISource, SqlServerSource, SqlSource, SquareSource, SybaseSource, TeradataSource, VerticaSource, XeroSource, ZohoSource. + sub-classes are: AmazonMWSSource, AmazonRdsForSqlServerSource, AmazonRedshiftSource, AzureMariaDBSource, AzureMySqlSource, AzurePostgreSqlSource, AzureSqlSource, AzureTableSource, CassandraSource, ConcurSource, CouchbaseSource, Db2Source, DrillSource, DynamicsAXSource, EloquaSource, GoogleAdWordsSource, GoogleBigQuerySource, GreenplumSource, HBaseSource, HiveSource, HubspotSource, ImpalaSource, InformixSource, JiraSource, MagentoSource, MariaDBSource, MarketoSource, MySqlSource, NetezzaSource, OdbcSource, OracleServiceCloudSource, PaypalSource, PhoenixSource, PostgreSqlSource, PrestoSource, QuickBooksSource, ResponsysSource, SalesforceMarketingCloudSource, SalesforceSource, SapBwSource, SapCloudForCustomerSource, SapEccSource, SapHanaSource, SapOpenHubSource, SapTableSource, ServiceNowSource, ShopifySource, SparkSource, SqlDWSource, SqlMISource, SqlServerSource, SqlSource, SquareSource, SybaseSource, TeradataSource, VerticaSource, XeroSource, ZohoSource. All required parameters must be populated in order to send to Azure. @@ -769,7 +760,7 @@ class TabularSource(CopySource): } _subtype_map = { - 'type': {'AmazonMWSSource': 'AmazonMWSSource', 'AmazonRedshiftSource': 'AmazonRedshiftSource', 'AzureMariaDBSource': 'AzureMariaDBSource', 'AzureMySqlSource': 'AzureMySqlSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AzureSqlSource': 'AzureSqlSource', 'AzureTableSource': 'AzureTableSource', 'CassandraSource': 'CassandraSource', 'ConcurSource': 'ConcurSource', 'CouchbaseSource': 'CouchbaseSource', 'Db2Source': 'Db2Source', 'DrillSource': 'DrillSource', 'DynamicsAXSource': 'DynamicsAXSource', 'EloquaSource': 'EloquaSource', 'GoogleAdWordsSource': 'GoogleAdWordsSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'GreenplumSource': 'GreenplumSource', 'HBaseSource': 'HBaseSource', 'HiveSource': 'HiveSource', 'HubspotSource': 'HubspotSource', 'ImpalaSource': 'ImpalaSource', 'InformixSource': 'InformixSource', 'JiraSource': 'JiraSource', 'MagentoSource': 'MagentoSource', 'MariaDBSource': 'MariaDBSource', 'MarketoSource': 'MarketoSource', 'MySqlSource': 'MySqlSource', 'NetezzaSource': 'NetezzaSource', 'OdbcSource': 'OdbcSource', 'OracleServiceCloudSource': 'OracleServiceCloudSource', 'PaypalSource': 'PaypalSource', 'PhoenixSource': 'PhoenixSource', 'PostgreSqlSource': 'PostgreSqlSource', 'PrestoSource': 'PrestoSource', 'QuickBooksSource': 'QuickBooksSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'SalesforceSource': 'SalesforceSource', 'SapBwSource': 'SapBwSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SapEccSource': 'SapEccSource', 'SapHanaSource': 'SapHanaSource', 'SapOpenHubSource': 'SapOpenHubSource', 'SapTableSource': 'SapTableSource', 'ServiceNowSource': 'ServiceNowSource', 'ShopifySource': 'ShopifySource', 'SparkSource': 'SparkSource', 'SqlDWSource': 'SqlDWSource', 'SqlMISource': 'SqlMISource', 'SqlServerSource': 'SqlServerSource', 'SqlSource': 'SqlSource', 'SquareSource': 'SquareSource', 'SybaseSource': 'SybaseSource', 'TeradataSource': 'TeradataSource', 'VerticaSource': 'VerticaSource', 'XeroSource': 'XeroSource', 'ZohoSource': 'ZohoSource'} + 'type': {'AmazonMWSSource': 'AmazonMWSSource', 'AmazonRdsForSqlServerSource': 'AmazonRdsForSqlServerSource', 'AmazonRedshiftSource': 'AmazonRedshiftSource', 'AzureMariaDBSource': 'AzureMariaDBSource', 'AzureMySqlSource': 'AzureMySqlSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AzureSqlSource': 'AzureSqlSource', 'AzureTableSource': 'AzureTableSource', 'CassandraSource': 'CassandraSource', 'ConcurSource': 'ConcurSource', 'CouchbaseSource': 'CouchbaseSource', 'Db2Source': 'Db2Source', 'DrillSource': 'DrillSource', 'DynamicsAXSource': 'DynamicsAXSource', 'EloquaSource': 'EloquaSource', 'GoogleAdWordsSource': 'GoogleAdWordsSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'GreenplumSource': 'GreenplumSource', 'HBaseSource': 'HBaseSource', 'HiveSource': 'HiveSource', 'HubspotSource': 'HubspotSource', 'ImpalaSource': 'ImpalaSource', 'InformixSource': 'InformixSource', 'JiraSource': 'JiraSource', 'MagentoSource': 'MagentoSource', 'MariaDBSource': 'MariaDBSource', 'MarketoSource': 'MarketoSource', 'MySqlSource': 'MySqlSource', 'NetezzaSource': 'NetezzaSource', 'OdbcSource': 'OdbcSource', 'OracleServiceCloudSource': 'OracleServiceCloudSource', 'PaypalSource': 'PaypalSource', 'PhoenixSource': 'PhoenixSource', 'PostgreSqlSource': 'PostgreSqlSource', 'PrestoSource': 'PrestoSource', 'QuickBooksSource': 'QuickBooksSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'SalesforceSource': 'SalesforceSource', 'SapBwSource': 'SapBwSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SapEccSource': 'SapEccSource', 'SapHanaSource': 'SapHanaSource', 'SapOpenHubSource': 'SapOpenHubSource', 'SapTableSource': 'SapTableSource', 'ServiceNowSource': 'ServiceNowSource', 'ShopifySource': 'ShopifySource', 'SparkSource': 'SparkSource', 'SqlDWSource': 'SqlDWSource', 'SqlMISource': 'SqlMISource', 'SqlServerSource': 'SqlServerSource', 'SqlSource': 'SqlSource', 'SquareSource': 'SquareSource', 'SybaseSource': 'SybaseSource', 'TeradataSource': 'TeradataSource', 'VerticaSource': 'VerticaSource', 'XeroSource': 'XeroSource', 'ZohoSource': 'ZohoSource'} } def __init__( @@ -851,6 +842,502 @@ def __init__( self.query = query +class AmazonRdsForOracleLinkedService(LinkedService): + """AmazonRdsForOracle database. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are deserialized to this + collection. + :type additional_properties: dict[str, any] + :param type: Required. Type of linked service.Constant filled by server. + :type type: str + :param connect_via: The integration runtime reference. + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the linked service. + :type annotations: list[any] + :param connection_string: Required. The connection string. Type: string, SecureString or + AzureKeyVaultSecretReference. + :type connection_string: any + :param password: The Azure key vault secret reference of password in connection string. + :type password: ~azure.synapse.artifacts.models.SecretBase + :param encrypted_credential: The encrypted credential used for authentication. Credentials are + encrypted using the integration runtime credential manager. Type: string (or Expression with + resultType string). + :type encrypted_credential: any + """ + + _validation = { + 'type': {'required': True}, + 'connection_string': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__( + self, + *, + connection_string: Any, + additional_properties: Optional[Dict[str, Any]] = None, + connect_via: Optional["IntegrationRuntimeReference"] = None, + description: Optional[str] = None, + parameters: Optional[Dict[str, "ParameterSpecification"]] = None, + annotations: Optional[List[Any]] = None, + password: Optional["SecretBase"] = None, + encrypted_credential: Optional[Any] = None, + **kwargs + ): + super(AmazonRdsForOracleLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'AmazonRdsForOracle' # type: str + self.connection_string = connection_string + self.password = password + self.encrypted_credential = encrypted_credential + + +class AmazonRdsForOraclePartitionSettings(msrest.serialization.Model): + """The settings that will be leveraged for AmazonRdsForOracle source partitioning. + + :param partition_names: Names of the physical partitions of AmazonRdsForOracle table. + :type partition_names: any + :param partition_column_name: The name of the column in integer type that will be used for + proceeding range partitioning. Type: string (or Expression with resultType string). + :type partition_column_name: any + :param partition_upper_bound: The maximum value of column specified in partitionColumnName that + will be used for proceeding range partitioning. Type: string (or Expression with resultType + string). + :type partition_upper_bound: any + :param partition_lower_bound: The minimum value of column specified in partitionColumnName that + will be used for proceeding range partitioning. Type: string (or Expression with resultType + string). + :type partition_lower_bound: any + """ + + _attribute_map = { + 'partition_names': {'key': 'partitionNames', 'type': 'object'}, + 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, + 'partition_upper_bound': {'key': 'partitionUpperBound', 'type': 'object'}, + 'partition_lower_bound': {'key': 'partitionLowerBound', 'type': 'object'}, + } + + def __init__( + self, + *, + partition_names: Optional[Any] = None, + partition_column_name: Optional[Any] = None, + partition_upper_bound: Optional[Any] = None, + partition_lower_bound: Optional[Any] = None, + **kwargs + ): + super(AmazonRdsForOraclePartitionSettings, self).__init__(**kwargs) + self.partition_names = partition_names + self.partition_column_name = partition_column_name + self.partition_upper_bound = partition_upper_bound + self.partition_lower_bound = partition_lower_bound + + +class AmazonRdsForOracleSource(CopySource): + """A copy activity AmazonRdsForOracle source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are deserialized to this + collection. + :type additional_properties: dict[str, any] + :param type: Required. Copy source type.Constant filled by server. + :type type: str + :param source_retry_count: Source retry count. Type: integer (or Expression with resultType + integer). + :type source_retry_count: any + :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: any + :param max_concurrent_connections: The maximum concurrent connection count for the source data + store. Type: integer (or Expression with resultType integer). + :type max_concurrent_connections: any + :param oracle_reader_query: AmazonRdsForOracle reader query. Type: string (or Expression with + resultType string). + :type oracle_reader_query: any + :param query_timeout: Query timeout. Type: string (or Expression with resultType string), + pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type query_timeout: any + :param partition_option: The partition mechanism that will be used for AmazonRdsForOracle read + in parallel. Type: string (or Expression with resultType string). + :type partition_option: any + :param partition_settings: The settings that will be leveraged for AmazonRdsForOracle source + partitioning. + :type partition_settings: ~azure.synapse.artifacts.models.AmazonRdsForOraclePartitionSettings + :param additional_columns: Specifies the additional columns to be added to source data. Type: + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, + 'oracle_reader_query': {'key': 'oracleReaderQuery', 'type': 'object'}, + 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, + 'partition_settings': {'key': 'partitionSettings', 'type': 'AmazonRdsForOraclePartitionSettings'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, + } + + def __init__( + self, + *, + additional_properties: Optional[Dict[str, Any]] = None, + source_retry_count: Optional[Any] = None, + source_retry_wait: Optional[Any] = None, + max_concurrent_connections: Optional[Any] = None, + oracle_reader_query: Optional[Any] = None, + query_timeout: Optional[Any] = None, + partition_option: Optional[Any] = None, + partition_settings: Optional["AmazonRdsForOraclePartitionSettings"] = None, + additional_columns: Optional[Any] = None, + **kwargs + ): + super(AmazonRdsForOracleSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) + self.type = 'AmazonRdsForOracleSource' # type: str + self.oracle_reader_query = oracle_reader_query + self.query_timeout = query_timeout + self.partition_option = partition_option + self.partition_settings = partition_settings + self.additional_columns = additional_columns + + +class AmazonRdsForOracleTableDataset(Dataset): + """The AmazonRdsForOracle database dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are deserialized to this + collection. + :type additional_properties: dict[str, any] + :param type: Required. Type of dataset.Constant filled by server. + :type type: str + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: array (or Expression + with resultType array), itemType: DatasetDataElement. + :type structure: any + :param schema: Columns that define the physical type schema of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetSchemaDataElement. + :type schema: any + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the Dataset. + :type annotations: list[any] + :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the + root level. + :type folder: ~azure.synapse.artifacts.models.DatasetFolder + :param schema_type_properties_schema: The schema name of the AmazonRdsForOracle database. Type: + string (or Expression with resultType string). + :type schema_type_properties_schema: any + :param table: The table name of the AmazonRdsForOracle database. Type: string (or Expression + with resultType string). + :type table: any + """ + + _validation = { + 'type': {'required': True}, + 'linked_service_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, + 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, + 'table': {'key': 'typeProperties.table', 'type': 'object'}, + } + + def __init__( + self, + *, + linked_service_name: "LinkedServiceReference", + additional_properties: Optional[Dict[str, Any]] = None, + description: Optional[str] = None, + structure: Optional[Any] = None, + schema: Optional[Any] = None, + parameters: Optional[Dict[str, "ParameterSpecification"]] = None, + annotations: Optional[List[Any]] = None, + folder: Optional["DatasetFolder"] = None, + schema_type_properties_schema: Optional[Any] = None, + table: Optional[Any] = None, + **kwargs + ): + super(AmazonRdsForOracleTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.type = 'AmazonRdsForOracleTable' # type: str + self.schema_type_properties_schema = schema_type_properties_schema + self.table = table + + +class AmazonRdsForSqlServerLinkedService(LinkedService): + """Amazon RDS for SQL Server linked service. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are deserialized to this + collection. + :type additional_properties: dict[str, any] + :param type: Required. Type of linked service.Constant filled by server. + :type type: str + :param connect_via: The integration runtime reference. + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param parameters: Parameters for linked service. + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the linked service. + :type annotations: list[any] + :param connection_string: Required. The connection string. Type: string, SecureString or + AzureKeyVaultSecretReference. + :type connection_string: any + :param user_name: The on-premises Windows authentication user name. Type: string (or Expression + with resultType string). + :type user_name: any + :param password: The on-premises Windows authentication password. + :type password: ~azure.synapse.artifacts.models.SecretBase + :param encrypted_credential: The encrypted credential used for authentication. Credentials are + encrypted using the integration runtime credential manager. Type: string (or Expression with + resultType string). + :type encrypted_credential: any + """ + + _validation = { + 'type': {'required': True}, + 'connection_string': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, + 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__( + self, + *, + connection_string: Any, + additional_properties: Optional[Dict[str, Any]] = None, + connect_via: Optional["IntegrationRuntimeReference"] = None, + description: Optional[str] = None, + parameters: Optional[Dict[str, "ParameterSpecification"]] = None, + annotations: Optional[List[Any]] = None, + user_name: Optional[Any] = None, + password: Optional["SecretBase"] = None, + encrypted_credential: Optional[Any] = None, + **kwargs + ): + super(AmazonRdsForSqlServerLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) + self.type = 'AmazonRdsForSqlServer' # type: str + self.connection_string = connection_string + self.user_name = user_name + self.password = password + self.encrypted_credential = encrypted_credential + + +class AmazonRdsForSqlServerSource(TabularSource): + """A copy activity Amazon RDS for SQL Server source. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are deserialized to this + collection. + :type additional_properties: dict[str, any] + :param type: Required. Copy source type.Constant filled by server. + :type type: str + :param source_retry_count: Source retry count. Type: integer (or Expression with resultType + integer). + :type source_retry_count: any + :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: any + :param max_concurrent_connections: The maximum concurrent connection count for the source data + store. Type: integer (or Expression with resultType integer). + :type max_concurrent_connections: any + :param query_timeout: Query timeout. Type: string (or Expression with resultType string), + pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type query_timeout: any + :param additional_columns: Specifies the additional columns to be added to source data. Type: + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any + :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). + :type sql_reader_query: any + :param sql_reader_stored_procedure_name: Name of the stored procedure for a SQL Database + source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression + with resultType string). + :type sql_reader_stored_procedure_name: any + :param stored_procedure_parameters: Value and type setting for stored procedure parameters. + Example: "{Parameter1: {value: "1", type: "int"}}". + :type stored_procedure_parameters: dict[str, + ~azure.synapse.artifacts.models.StoredProcedureParameter] + :param produce_additional_types: Which additional types to produce. + :type produce_additional_types: any + :param partition_option: The partition mechanism that will be used for Sql read in parallel. + Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". + :type partition_option: any + :param partition_settings: The settings that will be leveraged for Sql source partitioning. + :type partition_settings: ~azure.synapse.artifacts.models.SqlPartitionSettings + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, + 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, + 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, + 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, + 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, + 'produce_additional_types': {'key': 'produceAdditionalTypes', 'type': 'object'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, + 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, + } + + def __init__( + self, + *, + additional_properties: Optional[Dict[str, Any]] = None, + source_retry_count: Optional[Any] = None, + source_retry_wait: Optional[Any] = None, + max_concurrent_connections: Optional[Any] = None, + query_timeout: Optional[Any] = None, + additional_columns: Optional[Any] = None, + sql_reader_query: Optional[Any] = None, + sql_reader_stored_procedure_name: Optional[Any] = None, + stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, + produce_additional_types: Optional[Any] = None, + partition_option: Optional[Any] = None, + partition_settings: Optional["SqlPartitionSettings"] = None, + **kwargs + ): + super(AmazonRdsForSqlServerSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) + self.type = 'AmazonRdsForSqlServerSource' # type: str + self.sql_reader_query = sql_reader_query + self.sql_reader_stored_procedure_name = sql_reader_stored_procedure_name + self.stored_procedure_parameters = stored_procedure_parameters + self.produce_additional_types = produce_additional_types + self.partition_option = partition_option + self.partition_settings = partition_settings + + +class AmazonRdsForSqlServerTableDataset(Dataset): + """The Amazon RDS for SQL Server dataset. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are deserialized to this + collection. + :type additional_properties: dict[str, any] + :param type: Required. Type of dataset.Constant filled by server. + :type type: str + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: array (or Expression + with resultType array), itemType: DatasetDataElement. + :type structure: any + :param schema: Columns that define the physical type schema of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetSchemaDataElement. + :type schema: any + :param linked_service_name: Required. Linked service reference. + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] + :param annotations: List of tags that can be used for describing the Dataset. + :type annotations: list[any] + :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the + root level. + :type folder: ~azure.synapse.artifacts.models.DatasetFolder + :param schema_type_properties_schema: The schema name of the SQL Server dataset. Type: string + (or Expression with resultType string). + :type schema_type_properties_schema: any + :param table: The table name of the SQL Server dataset. Type: string (or Expression with + resultType string). + :type table: any + """ + + _validation = { + 'type': {'required': True}, + 'linked_service_name': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'schema': {'key': 'schema', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, + 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, + 'table': {'key': 'typeProperties.table', 'type': 'object'}, + } + + def __init__( + self, + *, + linked_service_name: "LinkedServiceReference", + additional_properties: Optional[Dict[str, Any]] = None, + description: Optional[str] = None, + structure: Optional[Any] = None, + schema: Optional[Any] = None, + parameters: Optional[Dict[str, "ParameterSpecification"]] = None, + annotations: Optional[List[Any]] = None, + folder: Optional["DatasetFolder"] = None, + schema_type_properties_schema: Optional[Any] = None, + table: Optional[Any] = None, + **kwargs + ): + super(AmazonRdsForSqlServerTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) + self.type = 'AmazonRdsForSqlServerTable' # type: str + self.schema_type_properties_schema = schema_type_properties_schema + self.table = table + + class AmazonRedshiftLinkedService(LinkedService): """Linked service for Amazon Redshift. @@ -862,13 +1349,11 @@ class AmazonRedshiftLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param server: Required. The name of the Amazon Redshift server. Type: string (or Expression @@ -878,7 +1363,7 @@ class AmazonRedshiftLinkedService(LinkedService): resultType string). :type username: any :param password: The password of the Amazon Redshift source. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param database: Required. The database name of the Amazon Redshift source. Type: string (or Expression with resultType string). :type database: any @@ -968,8 +1453,7 @@ class AmazonRedshiftSource(TabularSource): :param redshift_unload_settings: The Amazon S3 settings needed for the interim Amazon S3 when copying from Amazon Redshift with unload. With this, data from Amazon Redshift source will be unloaded into S3 first and then copied into the targeted sink from the interim S3. - :type redshift_unload_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.RedshiftUnloadSettings + :type redshift_unload_settings: ~azure.synapse.artifacts.models.RedshiftUnloadSettings """ _validation = { @@ -1026,16 +1510,14 @@ class AmazonRedshiftTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -1109,16 +1591,14 @@ class AmazonS3Dataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param bucket_name: Required. The name of the Amazon S3 bucket. Type: string (or Expression with resultType string). :type bucket_name: any @@ -1138,9 +1618,9 @@ class AmazonS3Dataset(Dataset): Expression with resultType string). :type modified_datetime_end: any :param format: The format of files. - :type format: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetStorageFormat + :type format: ~azure.synapse.artifacts.models.DatasetStorageFormat :param compression: The data compression method used for the Amazon S3 object. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression """ _validation = { @@ -1213,13 +1693,11 @@ class AmazonS3LinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param authentication_type: The authentication type of S3. Allowed value: AccessKey (default) @@ -1230,13 +1708,13 @@ class AmazonS3LinkedService(LinkedService): :type access_key_id: any :param secret_access_key: The secret access key of the Amazon S3 Identity and Access Management (IAM) user. - :type secret_access_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type secret_access_key: ~azure.synapse.artifacts.models.SecretBase :param service_url: This value specifies the endpoint to access with the S3 Connector. This is an optional property; change it only if you want to try a different service endpoint or want to switch between https and http. Type: string (or Expression with resultType string). :type service_url: any :param session_token: The session token for the S3 temporary security credential. - :type session_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type session_token: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -1551,9 +2029,9 @@ class ControlActivity(Activity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] """ _validation = { @@ -1603,9 +2081,9 @@ class AppendVariableActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param variable_name: Name of the variable whose value needs to be appended to. :type variable_name: str :param value: Value to be appended. Could be a static value or Expression. @@ -1748,18 +2226,16 @@ class AvroDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param location: The location of the avro storage. - :type location: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetLocation + :type location: ~azure.synapse.artifacts.models.DatasetLocation :param avro_compression_codec: A string from AvroCompressionCodecEnum or an expression. :type avro_compression_codec: any :param avro_compression_level: @@ -1993,9 +2469,9 @@ class AvroSink(CopySink): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Avro store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreWriteSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreWriteSettings :param format_settings: Avro format settings. - :type format_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.AvroWriteSettings + :type format_settings: ~azure.synapse.artifacts.models.AvroWriteSettings """ _validation = { @@ -2053,7 +2529,7 @@ class AvroSource(CopySource): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Avro store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :type additional_columns: any @@ -2194,20 +2670,18 @@ class AzureBatchLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param account_name: Required. The Azure Batch account name. Type: string (or Expression with resultType string). :type account_name: any :param access_key: The Azure Batch account access key. - :type access_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type access_key: ~azure.synapse.artifacts.models.SecretBase :param batch_uri: Required. The Azure Batch URI. Type: string (or Expression with resultType string). :type batch_uri: any @@ -2215,8 +2689,7 @@ class AzureBatchLinkedService(LinkedService): resultType string). :type pool_name: any :param linked_service_name: Required. The Azure Storage linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -2291,16 +2764,14 @@ class AzureBlobDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param folder_path: The path of the Azure Blob storage. Type: string (or Expression with resultType string). :type folder_path: any @@ -2317,9 +2788,9 @@ class AzureBlobDataset(Dataset): Expression with resultType string). :type modified_datetime_end: any :param format: The format of the Azure Blob storage. - :type format: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetStorageFormat + :type format: ~azure.synapse.artifacts.models.DatasetStorageFormat :param compression: The data compression method used for the blob storage. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression """ _validation = { @@ -2396,16 +2867,14 @@ class AzureBlobFSDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param folder_path: The path of the Azure Data Lake Storage Gen2 storage. Type: string (or Expression with resultType string). :type folder_path: any @@ -2413,9 +2882,9 @@ class AzureBlobFSDataset(Dataset): with resultType string). :type file_name: any :param format: The format of the Azure Data Lake Storage Gen2 storage. - :type format: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetStorageFormat + :type format: ~azure.synapse.artifacts.models.DatasetStorageFormat :param compression: The data compression method used for the blob storage. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression """ _validation = { @@ -2475,13 +2944,11 @@ class AzureBlobFSLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param url: Required. Endpoint for the Azure Data Lake Storage Gen2 service. Type: string (or @@ -2495,7 +2962,7 @@ class AzureBlobFSLinkedService(LinkedService): :type service_principal_id: any :param service_principal_key: The Key of the application used to authenticate against the Azure Data Lake Storage Gen2 account. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -2927,28 +3394,24 @@ class AzureBlobStorageLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: The connection string. It is mutually exclusive with sasUri, serviceEndpoint property. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param account_key: The Azure key vault secret reference of accountKey in connection string. - :type account_key: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type account_key: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param sas_uri: SAS URI of the Azure Blob Storage resource. It is mutually exclusive with connectionString, serviceEndpoint property. Type: string, SecureString or AzureKeyVaultSecretReference. :type sas_uri: any :param sas_token: The Azure key vault secret reference of sasToken in sas uri. - :type sas_token: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type sas_token: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param service_endpoint: Blob service endpoint of the Azure Blob Storage resource. It is mutually exclusive with connectionString, sasUri property. :type service_endpoint: str @@ -2957,7 +3420,7 @@ class AzureBlobStorageLinkedService(LinkedService): :type service_principal_id: any :param service_principal_key: The key of the service principal used to authenticate against Azure SQL Data Warehouse. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -2965,6 +3428,10 @@ class AzureBlobStorageLinkedService(LinkedService): values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data factory regions’ cloud type. Type: string (or Expression with resultType string). :type azure_cloud_type: any + :param account_kind: Specify the kind of your storage account. Allowed values are: Storage + (general purpose v1), StorageV2 (general purpose v2), BlobStorage, or BlockBlobStorage. Type: + string (or Expression with resultType string). + :type account_kind: str :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -2991,6 +3458,7 @@ class AzureBlobStorageLinkedService(LinkedService): 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, 'azure_cloud_type': {'key': 'typeProperties.azureCloudType', 'type': 'object'}, + 'account_kind': {'key': 'typeProperties.accountKind', 'type': 'str'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, } @@ -3011,6 +3479,7 @@ def __init__( service_principal_key: Optional["SecretBase"] = None, tenant: Optional[Any] = None, azure_cloud_type: Optional[Any] = None, + account_kind: Optional[str] = None, encrypted_credential: Optional[str] = None, **kwargs ): @@ -3025,6 +3494,7 @@ def __init__( self.service_principal_key = service_principal_key self.tenant = tenant self.azure_cloud_type = azure_cloud_type + self.account_kind = account_kind self.encrypted_credential = encrypted_credential @@ -3236,16 +3706,14 @@ class AzureDatabricksDeltaLakeDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table: The name of delta table. Type: string (or Expression with resultType string). :type table: any :param database: The database name of delta table. Type: string (or Expression with resultType @@ -3468,13 +3936,11 @@ class AzureDatabricksDeltaLakeLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param domain: Required. :code:``.azuredatabricks.net, domain name of your Databricks @@ -3483,7 +3949,7 @@ class AzureDatabricksDeltaLakeLinkedService(LinkedService): :param access_token: Required. Access token for databricks REST API. Refer to https://docs.azuredatabricks.net/api/latest/authentication.html. Type: string, SecureString or AzureKeyVaultSecretReference. - :type access_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type access_token: ~azure.synapse.artifacts.models.SecretBase :param cluster_id: The id of an existing interactive cluster that will be used for all runs of this job. Type: string (or Expression with resultType string). :type cluster_id: any @@ -3563,8 +4029,7 @@ class AzureDatabricksDeltaLakeSink(CopySink): string). :type pre_copy_script: any :param import_settings: Azure Databricks Delta Lake import settings. - :type import_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureDatabricksDeltaLakeImportCommand + :type import_settings: ~azure.synapse.artifacts.models.AzureDatabricksDeltaLakeImportCommand """ _validation = { @@ -3625,8 +4090,7 @@ class AzureDatabricksDeltaLakeSource(CopySource): resultType string). :type query: any :param export_settings: Azure Databricks Delta Lake export settings. - :type export_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureDatabricksDeltaLakeExportCommand + :type export_settings: ~azure.synapse.artifacts.models.AzureDatabricksDeltaLakeExportCommand """ _validation = { @@ -3671,13 +4135,11 @@ class AzureDatabricksLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param domain: Required. :code:``.azuredatabricks.net, domain name of your Databricks @@ -3686,7 +4148,7 @@ class AzureDatabricksLinkedService(LinkedService): :param access_token: Access token for databricks REST API. Refer to https://docs.azuredatabricks.net/api/latest/authentication.html. Type: string (or Expression with resultType string). - :type access_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type access_token: ~azure.synapse.artifacts.models.SecretBase :param authentication: Required to specify MSI, if using Workspace resource id for databricks REST API. Type: string (or Expression with resultType string). :type authentication: any @@ -3847,14 +4309,13 @@ class ExecutionActivity(Activity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy """ _validation = { @@ -3910,14 +4371,13 @@ class AzureDataExplorerCommandActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param command: Required. A control command, according to the Azure Data Explorer command syntax. Type: string (or Expression with resultType string). :type command: any @@ -3976,40 +4436,35 @@ class AzureDataExplorerLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param endpoint: Required. The endpoint of Azure Data Explorer (the engine's endpoint). URL will be in the format https://:code:``.:code:``.kusto.windows.net. Type: string (or Expression with resultType string). :type endpoint: any - :param service_principal_id: Required. The ID of the service principal used to authenticate - against Azure Data Explorer. Type: string (or Expression with resultType string). + :param service_principal_id: The ID of the service principal used to authenticate against Azure + Data Explorer. Type: string (or Expression with resultType string). :type service_principal_id: any - :param service_principal_key: Required. The key of the service principal used to authenticate - against Kusto. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :param service_principal_key: The key of the service principal used to authenticate against + Kusto. + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param database: Required. Database name for connection. Type: string (or Expression with resultType string). :type database: any - :param tenant: Required. The name or ID of the tenant to which the service principal belongs. - Type: string (or Expression with resultType string). + :param tenant: The name or ID of the tenant to which the service principal belongs. Type: + string (or Expression with resultType string). :type tenant: any """ _validation = { 'type': {'required': True}, 'endpoint': {'required': True}, - 'service_principal_id': {'required': True}, - 'service_principal_key': {'required': True}, 'database': {'required': True}, - 'tenant': {'required': True}, } _attribute_map = { @@ -4030,15 +4485,15 @@ def __init__( self, *, endpoint: Any, - service_principal_id: Any, - service_principal_key: "SecretBase", database: Any, - tenant: Any, additional_properties: Optional[Dict[str, Any]] = None, connect_via: Optional["IntegrationRuntimeReference"] = None, description: Optional[str] = None, parameters: Optional[Dict[str, "ParameterSpecification"]] = None, annotations: Optional[List[Any]] = None, + service_principal_id: Optional[Any] = None, + service_principal_key: Optional["SecretBase"] = None, + tenant: Optional[Any] = None, **kwargs ): super(AzureDataExplorerLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) @@ -4214,16 +4669,14 @@ class AzureDataExplorerTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table: The table name of the Azure Data Explorer database. Type: string (or Expression with resultType string). :type table: any @@ -4277,13 +4730,11 @@ class AzureDataLakeAnalyticsLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param account_name: Required. The Azure Data Lake Analytics account name. Type: string (or @@ -4294,7 +4745,7 @@ class AzureDataLakeAnalyticsLinkedService(LinkedService): :type service_principal_id: any :param service_principal_key: The Key of the application used to authenticate against the Azure Data Lake Analytics account. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: Required. The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -4385,16 +4836,14 @@ class AzureDataLakeStoreDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param folder_path: Path to the folder in the Azure Data Lake Store. Type: string (or Expression with resultType string). :type folder_path: any @@ -4402,10 +4851,10 @@ class AzureDataLakeStoreDataset(Dataset): Expression with resultType string). :type file_name: any :param format: The format of the Data Lake Store. - :type format: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetStorageFormat + :type format: ~azure.synapse.artifacts.models.DatasetStorageFormat :param compression: The data compression method used for the item(s) in the Azure Data Lake Store. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression """ _validation = { @@ -4465,13 +4914,11 @@ class AzureDataLakeStoreLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param data_lake_store_uri: Required. Data Lake Store service URI. Type: string (or Expression @@ -4482,7 +4929,7 @@ class AzureDataLakeStoreLinkedService(LinkedService): :type service_principal_id: any :param service_principal_key: The Key of the application used to authenticate against the Azure Data Lake Store account. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -4957,13 +5404,11 @@ class AzureFileStorageLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. Host name of the server. Type: string (or Expression with resultType @@ -4973,19 +5418,17 @@ class AzureFileStorageLinkedService(LinkedService): string). :type user_id: any :param password: Password to logon the server. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param connection_string: The connection string. It is mutually exclusive with sasUri property. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param account_key: The Azure key vault secret reference of accountKey in connection string. - :type account_key: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type account_key: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param sas_uri: SAS URI of the Azure File resource. It is mutually exclusive with connectionString property. Type: string, SecureString or AzureKeyVaultSecretReference. :type sas_uri: any :param sas_token: The Azure key vault secret reference of sasToken in sas uri. - :type sas_token: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type sas_token: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param file_share: The azure file share name. It is required when auth with accountKey/sasToken. Type: string (or Expression with resultType string). :type file_share: any @@ -5248,18 +5691,16 @@ class AzureFunctionActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param method: Required. Rest API method for target endpoint. Possible values include: "GET", "POST", "PUT", "DELETE", "OPTIONS", "HEAD", "TRACE". - :type method: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureFunctionActivityMethod + :type method: str or ~azure.synapse.artifacts.models.AzureFunctionActivityMethod :param function_name: Required. Name of the Function that the Azure Function Activity will call. Type: string (or Expression with resultType string). :type function_name: any @@ -5329,20 +5770,18 @@ class AzureFunctionLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param function_app_url: Required. The endpoint of the Azure Function App. URL will be in the format https://:code:``.azurewebsites.net. :type function_app_url: any :param function_key: Function or Host key for Azure Function App. - :type function_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type function_key: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5397,13 +5836,11 @@ class AzureKeyVaultLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param base_url: Required. The base URL of the Azure Key Vault. e.g. @@ -5482,7 +5919,7 @@ class AzureKeyVaultSecretReference(SecretBase): :param type: Required. Type of the secret.Constant filled by server. :type type: str :param store: Required. The Azure Key Vault linked service reference. - :type store: ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type store: ~azure.synapse.artifacts.models.LinkedServiceReference :param secret_name: Required. The name of the secret in Azure Key Vault. Type: string (or Expression with resultType string). :type secret_name: any @@ -5530,20 +5967,18 @@ class AzureMariaDBLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type pwd: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5667,16 +6102,14 @@ class AzureMariaDBTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -5733,14 +6166,13 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param global_parameters: Key,Value pairs to be passed to the Azure ML Batch Execution Service endpoint. Keys must match the names of web service parameters defined in the published Azure ML web service. Values will be passed in the GlobalParameters property of the Azure ML batch @@ -5750,14 +6182,12 @@ class AzureMLBatchExecutionActivity(ExecutionActivity): Service Outputs to AzureMLWebServiceFile objects specifying the output Blob locations. This information will be passed in the WebServiceOutputs property of the Azure ML batch execution request. - :type web_service_outputs: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureMLWebServiceFile] + :type web_service_outputs: dict[str, ~azure.synapse.artifacts.models.AzureMLWebServiceFile] :param web_service_inputs: Key,Value pairs, mapping the names of Azure ML endpoint's Web Service Inputs to AzureMLWebServiceFile objects specifying the input Blob locations.. This information will be passed in the WebServiceInputs property of the Azure ML batch execution request. - :type web_service_inputs: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureMLWebServiceFile] + :type web_service_inputs: dict[str, ~azure.synapse.artifacts.models.AzureMLWebServiceFile] """ _validation = { @@ -5816,14 +6246,13 @@ class AzureMLExecutePipelineActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param ml_pipeline_id: Required. ID of the published Azure ML pipeline. Type: string (or Expression with resultType string). :type ml_pipeline_id: any @@ -5906,20 +6335,18 @@ class AzureMLLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param ml_endpoint: Required. The Batch Execution REST URL for an Azure ML Studio Web Service endpoint. Type: string (or Expression with resultType string). :type ml_endpoint: any :param api_key: Required. The API key for accessing the Azure ML model endpoint. - :type api_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type api_key: ~azure.synapse.artifacts.models.SecretBase :param update_resource_endpoint: The Update Resource REST URL for an Azure ML Studio Web Service endpoint. Type: string (or Expression with resultType string). :type update_resource_endpoint: any @@ -5929,7 +6356,7 @@ class AzureMLLinkedService(LinkedService): :type service_principal_id: any :param service_principal_key: The key of the service principal used to authenticate against the ARM-based updateResourceEndpoint of an Azure ML Studio web service. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -6000,13 +6427,11 @@ class AzureMLServiceLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param subscription_id: Required. Azure ML Service workspace subscription ID. Type: string (or @@ -6024,7 +6449,7 @@ class AzureMLServiceLinkedService(LinkedService): :type service_principal_id: any :param service_principal_key: The key of the service principal used to authenticate against the endpoint of a published Azure ML Service pipeline. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -6100,21 +6525,19 @@ class AzureMLUpdateResourceActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param trained_model_name: Required. Name of the Trained Model module in the Web Service experiment to be updated. Type: string (or Expression with resultType string). :type trained_model_name: any :param trained_model_linked_service_name: Required. Name of Azure Storage linked service holding the .ilearner file that will be uploaded by the update operation. - :type trained_model_linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type trained_model_linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param trained_model_file_path: Required. The relative file path in trainedModelLinkedService to represent the .ilearner file that will be uploaded by the update operation. Type: string (or Expression with resultType string). @@ -6175,8 +6598,7 @@ class AzureMLWebServiceFile(msrest.serialization.Model): :type file_path: any :param linked_service_name: Required. Reference to an Azure Storage LinkedService, where Azure ML WebService Input/Output file located. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference """ _validation = { @@ -6212,21 +6634,18 @@ class AzureMySqlLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param password: The Azure key vault secret reference of password in connection string. - :type password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type password: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6412,16 +6831,14 @@ class AzureMySqlTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The Azure MySQL database table name. Type: string (or Expression with resultType string). :type table_name: any @@ -6481,21 +6898,18 @@ class AzurePostgreSqlLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param password: The Azure key vault secret reference of password in connection string. - :type password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type password: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6681,16 +7095,14 @@ class AzurePostgreSqlTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name of the Azure PostgreSQL database which includes both schema and table. Type: string (or Expression with resultType string). :type table_name: any @@ -6820,16 +7232,14 @@ class AzureSearchIndexDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param index_name: Required. The name of the Azure Search Index. Type: string (or Expression with resultType string). :type index_name: any @@ -6900,8 +7310,7 @@ class AzureSearchIndexSink(CopySink): :type max_concurrent_connections: any :param write_behavior: Specify the write behavior when upserting documents into Azure Search Index. Possible values include: "Merge", "Upload". - :type write_behavior: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureSearchIndexWriteBehaviorType + :type write_behavior: str or ~azure.synapse.artifacts.models.AzureSearchIndexWriteBehaviorType """ _validation = { @@ -6947,20 +7356,18 @@ class AzureSearchLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param url: Required. URL for Azure Search service. Type: string (or Expression with resultType string). :type url: any :param key: Admin Key for Azure Search service. - :type key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type key: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -7015,27 +7422,24 @@ class AzureSqlDatabaseLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param password: The Azure key vault secret reference of password in connection string. - :type password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type password: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param service_principal_id: The ID of the service principal used to authenticate against Azure SQL Database. Type: string (or Expression with resultType string). :type service_principal_id: any :param service_principal_key: The key of the service principal used to authenticate against Azure SQL Database. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -7109,27 +7513,24 @@ class AzureSqlDWLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param password: The Azure key vault secret reference of password in connection string. - :type password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type password: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param service_principal_id: The ID of the service principal used to authenticate against Azure SQL Data Warehouse. Type: string (or Expression with resultType string). :type service_principal_id: any :param service_principal_key: The key of the service principal used to authenticate against Azure SQL Data Warehouse. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -7211,16 +7612,14 @@ class AzureSqlDWTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -7286,27 +7685,24 @@ class AzureSqlMILinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param password: The Azure key vault secret reference of password in connection string. - :type password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type password: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param service_principal_id: The ID of the service principal used to authenticate against Azure SQL Managed Instance. Type: string (or Expression with resultType string). :type service_principal_id: any :param service_principal_key: The key of the service principal used to authenticate against Azure SQL Managed Instance. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -7388,16 +7784,14 @@ class AzureSqlMITableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -7488,7 +7882,7 @@ class AzureSqlSink(CopySink): :type pre_copy_script: any :param stored_procedure_parameters: SQL stored procedure parameters. :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] :param stored_procedure_table_type_parameter_name: The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). :type stored_procedure_table_type_parameter_name: any @@ -7578,15 +7972,14 @@ class AzureSqlSource(TabularSource): :param stored_procedure_parameters: Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] :param produce_additional_types: Which additional types to produce. :type produce_additional_types: any :param partition_option: The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". :type partition_option: any :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.SqlPartitionSettings """ _validation = { @@ -7655,16 +8048,14 @@ class AzureSqlTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -7730,27 +8121,23 @@ class AzureStorageLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: The connection string. It is mutually exclusive with sasUri property. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param account_key: The Azure key vault secret reference of accountKey in connection string. - :type account_key: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type account_key: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param sas_uri: SAS URI of the Azure Storage resource. It is mutually exclusive with connectionString property. Type: string, SecureString or AzureKeyVaultSecretReference. :type sas_uri: any :param sas_token: The Azure key vault secret reference of sasToken in sas uri. - :type sas_token: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type sas_token: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -7818,16 +8205,14 @@ class AzureTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: Required. The table name of the Azure Table storage. Type: string (or Expression with resultType string). :type table_name: any @@ -8030,27 +8415,23 @@ class AzureTableStorageLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: The connection string. It is mutually exclusive with sasUri property. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param account_key: The Azure key vault secret reference of accountKey in connection string. - :type account_key: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type account_key: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param sas_uri: SAS URI of the Azure Storage resource. It is mutually exclusive with connectionString property. Type: string, SecureString or AzureKeyVaultSecretReference. :type sas_uri: any :param sas_token: The Azure key vault secret reference of sasToken in sas uri. - :type sas_token: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type sas_token: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -8106,7 +8487,7 @@ class BigDataPoolReference(msrest.serialization.Model): :param type: Required. Big data pool reference type. Possible values include: "BigDataPoolReference". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.BigDataPoolReferenceType + :type type: str or ~azure.synapse.artifacts.models.BigDataPoolReferenceType :param reference_name: Required. Reference big data pool name. :type reference_name: str """ @@ -8203,11 +8584,11 @@ class BigDataPoolResourceInfo(TrackedResource): :param provisioning_state: The state of the Big Data pool. :type provisioning_state: str :param auto_scale: Auto-scaling properties. - :type auto_scale: ~azure.synapse.artifacts.v2021_06_01_preview.models.AutoScaleProperties + :type auto_scale: ~azure.synapse.artifacts.models.AutoScaleProperties :param creation_date: The time when the Big Data pool was created. :type creation_date: ~datetime.datetime :param auto_pause: Auto-pausing properties. - :type auto_pause: ~azure.synapse.artifacts.v2021_06_01_preview.models.AutoPauseProperties + :type auto_pause: ~azure.synapse.artifacts.models.AutoPauseProperties :param is_compute_isolation_enabled: Whether compute isolation is required or not. :type is_compute_isolation_enabled: bool :param session_level_packages_enabled: Whether session level packages enabled. @@ -8215,31 +8596,27 @@ class BigDataPoolResourceInfo(TrackedResource): :param cache_size: The cache size. :type cache_size: int :param dynamic_executor_allocation: Dynamic Executor Allocation. - :type dynamic_executor_allocation: - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicExecutorAllocation + :type dynamic_executor_allocation: ~azure.synapse.artifacts.models.DynamicExecutorAllocation :param spark_events_folder: The Spark events folder. :type spark_events_folder: str :param node_count: The number of nodes in the Big Data pool. :type node_count: int :param library_requirements: Library version requirements. - :type library_requirements: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LibraryRequirements + :type library_requirements: ~azure.synapse.artifacts.models.LibraryRequirements :param custom_libraries: List of custom libraries/packages associated with the spark pool. - :type custom_libraries: list[~azure.synapse.artifacts.v2021_06_01_preview.models.LibraryInfo] + :type custom_libraries: list[~azure.synapse.artifacts.models.LibraryInfo] :param spark_config_properties: Spark configuration file to specify additional properties. - :type spark_config_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LibraryRequirements + :type spark_config_properties: ~azure.synapse.artifacts.models.LibraryRequirements :param spark_version: The Apache Spark version. :type spark_version: str :param default_spark_log_folder: The default folder where Spark logs will be written. :type default_spark_log_folder: str :param node_size: The level of compute power that each node in the Big Data pool has. Possible values include: "None", "Small", "Medium", "Large", "XLarge", "XXLarge", "XXXLarge". - :type node_size: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.NodeSize + :type node_size: str or ~azure.synapse.artifacts.models.NodeSize :param node_size_family: The kind of nodes that the Big Data pool provides. Possible values include: "None", "MemoryOptimized". - :type node_size_family: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.NodeSizeFamily + :type node_size_family: str or ~azure.synapse.artifacts.models.NodeSizeFamily :ivar last_succeeded_timestamp: The time when the Big Data pool was updated successfully. :vartype last_succeeded_timestamp: ~datetime.datetime """ @@ -8329,7 +8706,7 @@ class BigDataPoolResourceInfoListResult(msrest.serialization.Model): :param next_link: Link to the next page of results. :type next_link: str :param value: List of Big Data pools. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.BigDataPoolResourceInfo] + :type value: list[~azure.synapse.artifacts.models.BigDataPoolResourceInfo] """ _attribute_map = { @@ -8368,20 +8745,18 @@ class BinaryDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param location: The location of the Binary storage. - :type location: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetLocation + :type location: ~azure.synapse.artifacts.models.DatasetLocation :param compression: The data compression method used for the binary dataset. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression """ _validation = { @@ -8474,8 +8849,7 @@ class BinaryReadSettings(FormatReadSettings): :param type: Required. The read setting type.Constant filled by server. :type type: str :param compression_properties: Compression settings. - :type compression_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.CompressionReadSettings + :type compression_properties: ~azure.synapse.artifacts.models.CompressionReadSettings """ _validation = { @@ -8526,7 +8900,7 @@ class BinarySink(CopySink): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Binary store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreWriteSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreWriteSettings """ _validation = { @@ -8581,9 +8955,9 @@ class BinarySource(CopySource): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Binary store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings :param format_settings: Binary format settings. - :type format_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.BinaryReadSettings + :type format_settings: ~azure.synapse.artifacts.models.BinaryReadSettings """ _validation = { @@ -8636,8 +9010,7 @@ class Trigger(msrest.serialization.Model): :type description: str :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRuntimeState + :vartype runtime_state: str or ~azure.synapse.artifacts.models.TriggerRuntimeState :param annotations: List of tags that can be used for describing the trigger. :type annotations: list[any] """ @@ -8694,13 +9067,11 @@ class MultiplePipelineTrigger(Trigger): :type description: str :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRuntimeState + :vartype runtime_state: str or ~azure.synapse.artifacts.models.TriggerRuntimeState :param annotations: List of tags that can be used for describing the trigger. :type annotations: list[any] :param pipelines: Pipelines that need to be started. - :type pipelines: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerPipelineReference] + :type pipelines: list[~azure.synapse.artifacts.models.TriggerPipelineReference] """ _validation = { @@ -8751,13 +9122,11 @@ class BlobEventsTrigger(MultiplePipelineTrigger): :type description: str :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRuntimeState + :vartype runtime_state: str or ~azure.synapse.artifacts.models.TriggerRuntimeState :param annotations: List of tags that can be used for describing the trigger. :type annotations: list[any] :param pipelines: Pipelines that need to be started. - :type pipelines: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerPipelineReference] + :type pipelines: list[~azure.synapse.artifacts.models.TriggerPipelineReference] :param blob_path_begins_with: The blob path must begin with the pattern provided for trigger to fire. For example, '/records/blobs/december/' will only fire the trigger for blobs in the december folder under the records container. At least one of these must be provided: @@ -8770,7 +9139,7 @@ class BlobEventsTrigger(MultiplePipelineTrigger): :param ignore_empty_blobs: If set to true, blobs with zero bytes will be ignored. :type ignore_empty_blobs: bool :param events: Required. The type of events that cause this trigger to fire. - :type events: list[str or ~azure.synapse.artifacts.v2021_06_01_preview.models.BlobEventType] + :type events: list[str or ~azure.synapse.artifacts.models.BlobEventType] :param scope: Required. The ARM resource ID of the Storage Account. :type scope: str """ @@ -8978,21 +9347,18 @@ class BlobTrigger(MultiplePipelineTrigger): :type description: str :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRuntimeState + :vartype runtime_state: str or ~azure.synapse.artifacts.models.TriggerRuntimeState :param annotations: List of tags that can be used for describing the trigger. :type annotations: list[any] :param pipelines: Pipelines that need to be started. - :type pipelines: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerPipelineReference] + :type pipelines: list[~azure.synapse.artifacts.models.TriggerPipelineReference] :param folder_path: Required. The path of the container/folder that will trigger the pipeline. :type folder_path: str :param max_concurrency: Required. The max number of parallel files to handle when it is triggered. :type max_concurrency: int :param linked_service: Required. The Azure Storage linked service reference. - :type linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference """ _validation = { @@ -9045,13 +9411,11 @@ class CassandraLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. Host name for connection. Type: string (or Expression with resultType @@ -9067,7 +9431,7 @@ class CassandraLinkedService(LinkedService): string). :type username: any :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -9155,7 +9519,7 @@ class CassandraSource(TabularSource): Possible values include: "ALL", "EACH_QUORUM", "QUORUM", "LOCAL_QUORUM", "ONE", "TWO", "THREE", "LOCAL_ONE", "SERIAL", "LOCAL_SERIAL". :type consistency_level: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.CassandraSourceReadConsistencyLevels + ~azure.synapse.artifacts.models.CassandraSourceReadConsistencyLevels """ _validation = { @@ -9212,16 +9576,14 @@ class CassandraTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name of the Cassandra database. Type: string (or Expression with resultType string). :type table_name: any @@ -9286,15 +9648,14 @@ class ChainingTrigger(Trigger): :type description: str :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRuntimeState + :vartype runtime_state: str or ~azure.synapse.artifacts.models.TriggerRuntimeState :param annotations: List of tags that can be used for describing the trigger. :type annotations: list[any] :param pipeline: Required. Pipeline for which runs are created when all upstream pipelines complete successfully. - :type pipeline: ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerPipelineReference + :type pipeline: ~azure.synapse.artifacts.models.TriggerPipelineReference :param depends_on: Required. Upstream Pipelines. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineReference] + :type depends_on: list[~azure.synapse.artifacts.models.PipelineReference] :param run_dimension: Required. Run Dimension property that needs to be emitted by upstream pipelines. :type run_dimension: str @@ -9349,7 +9710,7 @@ class CloudError(msrest.serialization.Model): :param target: Property name/path in request associated with error. :type target: str :param details: Array with additional error details. - :type details: list[~azure.synapse.artifacts.v2021_06_01_preview.models.CloudError] + :type details: list[~azure.synapse.artifacts.models.CloudError] """ _validation = { @@ -9380,6 +9741,49 @@ def __init__( self.details = details +class CloudErrorAutoGenerated(msrest.serialization.Model): + """The object that defines the structure of an Azure Synapse error response. + + All required parameters must be populated in order to send to Azure. + + :param code: Required. Error code. + :type code: str + :param message: Required. Error message. + :type message: str + :param target: Property name/path in request associated with error. + :type target: str + :param details: Array with additional error details. + :type details: list[~azure.synapse.artifacts.models.CloudErrorAutoGenerated] + """ + + _validation = { + 'code': {'required': True}, + 'message': {'required': True}, + } + + _attribute_map = { + 'code': {'key': 'error.code', 'type': 'str'}, + 'message': {'key': 'error.message', 'type': 'str'}, + 'target': {'key': 'error.target', 'type': 'str'}, + 'details': {'key': 'error.details', 'type': '[CloudErrorAutoGenerated]'}, + } + + def __init__( + self, + *, + code: str, + message: str, + target: Optional[str] = None, + details: Optional[List["CloudErrorAutoGenerated"]] = None, + **kwargs + ): + super(CloudErrorAutoGenerated, self).__init__(**kwargs) + self.code = code + self.message = message + self.target = target + self.details = details + + class CommonDataServiceForAppsEntityDataset(Dataset): """The Common Data Service for Apps entity dataset. @@ -9399,16 +9803,14 @@ class CommonDataServiceForAppsEntityDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param entity_name: The logical name of the entity. Type: string (or Expression with resultType string). :type entity_name: any @@ -9462,21 +9864,18 @@ class CommonDataServiceForAppsLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param deployment_type: Required. The deployment type of the Common Data Service for Apps instance. 'Online' for Common Data Service for Apps Online and 'OnPremisesWithIfd' for Common Data Service for Apps on-premises with Ifd. Type: string (or Expression with resultType - string). Possible values include: "Online", "OnPremisesWithIfd". - :type deployment_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsDeploymentType + string). + :type deployment_type: any :param host_name: The host name of the on-premises Common Data Service for Apps server. The property is required for on-prem and not allowed for online. Type: string (or Expression with resultType string). @@ -9497,15 +9896,13 @@ class CommonDataServiceForAppsLinkedService(LinkedService): :param authentication_type: Required. The authentication type to connect to Common Data Service for Apps server. 'Office365' for online scenario, 'Ifd' for on-premises with Ifd scenario. 'AADServicePrincipal' for Server-To-Server authentication in online scenario. Type: string (or - Expression with resultType string). Possible values include: "Office365", "Ifd", - "AADServicePrincipal". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsAuthenticationType + Expression with resultType string). + :type authentication_type: any :param username: User name to access the Common Data Service for Apps instance. Type: string (or Expression with resultType string). :type username: any :param password: Password to access the Common Data Service for Apps instance. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param service_principal_id: The client ID of the application in Azure Active Directory used for Server-To-Server authentication. Type: string (or Expression with resultType string). :type service_principal_id: any @@ -9517,8 +9914,7 @@ class CommonDataServiceForAppsLinkedService(LinkedService): servicePrincipalCredential can be SecureString or AzureKeyVaultSecretReference. If servicePrincipalCredentialType is 'ServicePrincipalCert', servicePrincipalCredential can only be AzureKeyVaultSecretReference. - :type service_principal_credential: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_credential: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -9538,12 +9934,12 @@ class CommonDataServiceForAppsLinkedService(LinkedService): 'description': {'key': 'description', 'type': 'str'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'str'}, + 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'object'}, 'host_name': {'key': 'typeProperties.hostName', 'type': 'object'}, 'port': {'key': 'typeProperties.port', 'type': 'object'}, 'service_uri': {'key': 'typeProperties.serviceUri', 'type': 'object'}, 'organization_name': {'key': 'typeProperties.organizationName', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, 'username': {'key': 'typeProperties.username', 'type': 'object'}, 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, @@ -9555,8 +9951,8 @@ class CommonDataServiceForAppsLinkedService(LinkedService): def __init__( self, *, - deployment_type: Union[str, "DynamicsDeploymentType"], - authentication_type: Union[str, "DynamicsAuthenticationType"], + deployment_type: Any, + authentication_type: Any, additional_properties: Optional[Dict[str, Any]] = None, connect_via: Optional["IntegrationRuntimeReference"] = None, description: Optional[str] = None, @@ -9617,8 +10013,7 @@ class CommonDataServiceForAppsSink(CopySink): :type max_concurrent_connections: any :param write_behavior: Required. The write behavior for the operation. Possible values include: "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsSinkWriteBehavior + :type write_behavior: str or ~azure.synapse.artifacts.models.DynamicsSinkWriteBehavior :param ignore_null_values: The flag indicating whether to ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). @@ -9775,13 +10170,11 @@ class ConcurLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_properties: Properties used to connect to Concur. It is mutually exclusive @@ -9793,7 +10186,7 @@ class ConcurLinkedService(LinkedService): :type username: any :param password: The password corresponding to the user name that you provided in the username field. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -9882,16 +10275,14 @@ class ConcurObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -10010,29 +10401,28 @@ class CopyActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param inputs: List of inputs for the activity. - :type inputs: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference] + :type inputs: list[~azure.synapse.artifacts.models.DatasetReference] :param outputs: List of outputs for the activity. - :type outputs: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference] + :type outputs: list[~azure.synapse.artifacts.models.DatasetReference] :param source: Required. Copy activity source. - :type source: ~azure.synapse.artifacts.v2021_06_01_preview.models.CopySource + :type source: ~azure.synapse.artifacts.models.CopySource :param sink: Required. Copy activity sink. - :type sink: ~azure.synapse.artifacts.v2021_06_01_preview.models.CopySink + :type sink: ~azure.synapse.artifacts.models.CopySink :param translator: Copy activity translator. If not specified, tabular translator is used. :type translator: any :param enable_staging: Specifies whether to copy data via an interim staging. Default value is false. Type: boolean (or Expression with resultType boolean). :type enable_staging: any :param staging_settings: Specifies interim staging settings when EnableStaging is true. - :type staging_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StagingSettings + :type staging_settings: ~azure.synapse.artifacts.models.StagingSettings :param parallel_copies: Maximum number of concurrent sessions opened on the source or sink to avoid overloading the data store. Type: integer (or Expression with resultType integer), minimum: 0. @@ -10046,13 +10436,12 @@ class CopyActivity(ExecutionActivity): :param redirect_incompatible_row_settings: Redirect incompatible row settings when EnableSkipIncompatibleRow is true. :type redirect_incompatible_row_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.RedirectIncompatibleRowSettings + ~azure.synapse.artifacts.models.RedirectIncompatibleRowSettings :param log_storage_settings: (Deprecated. Please use LogSettings) Log storage settings customer need to provide when enabling session log. - :type log_storage_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LogStorageSettings + :type log_storage_settings: ~azure.synapse.artifacts.models.LogStorageSettings :param log_settings: Log settings customer needs provide when enabling log. - :type log_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.LogSettings + :type log_settings: ~azure.synapse.artifacts.models.LogSettings :param preserve_rules: Preserve Rules. :type preserve_rules: list[any] :param preserve: Preserve rules. @@ -10061,7 +10450,7 @@ class CopyActivity(ExecutionActivity): (or Expression with resultType boolean). :type validate_data_consistency: any :param skip_error_file: Specify the fault tolerance for data consistency. - :type skip_error_file: ~azure.synapse.artifacts.v2021_06_01_preview.models.SkipErrorFile + :type skip_error_file: ~azure.synapse.artifacts.models.SkipErrorFile """ _validation = { @@ -10227,13 +10616,11 @@ class CosmosDbLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: The connection string. Type: string, SecureString or @@ -10246,7 +10633,7 @@ class CosmosDbLinkedService(LinkedService): :type database: any :param account_key: The account key of the Azure CosmosDB account. Type: SecureString or AzureKeyVaultSecretReference. - :type account_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type account_key: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -10314,16 +10701,14 @@ class CosmosDbMongoDbApiCollectionDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param collection: Required. The collection name of the CosmosDB (MongoDB API) database. Type: string (or Expression with resultType string). :type collection: any @@ -10378,13 +10763,11 @@ class CosmosDbMongoDbApiLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The CosmosDB (MongoDB API) connection string. Type: string, @@ -10518,8 +10901,7 @@ class CosmosDbMongoDbApiSource(CopySource): with resultType string). :type filter: any :param cursor_methods: Cursor methods for Mongodb query. - :type cursor_methods: - ~azure.synapse.artifacts.v2021_06_01_preview.models.MongoDbCursorMethodsProperties + :type cursor_methods: ~azure.synapse.artifacts.models.MongoDbCursorMethodsProperties :param batch_size: Specifies the number of documents to return in each batch of the response from MongoDB instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. @@ -10592,16 +10974,14 @@ class CosmosDbSqlApiCollectionDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param collection_name: Required. CosmosDB (SQL API) collection name. Type: string (or Expression with resultType string). :type collection_name: any @@ -10793,21 +11173,18 @@ class CouchbaseLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param cred_string: The Azure key vault secret reference of credString in connection string. - :type cred_string: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type cred_string: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -10931,16 +11308,14 @@ class CouchbaseTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -10985,48 +11360,39 @@ def __init__( class CreateDataFlowDebugSessionRequest(msrest.serialization.Model): """Request body structure for creating data flow debug session. - :param data_flow_name: The name of the data flow. - :type data_flow_name: str - :param existing_cluster_id: The ID of existing Databricks cluster. - :type existing_cluster_id: str - :param cluster_timeout: Timeout setting for Databricks cluster. - :type cluster_timeout: int - :param new_cluster_name: The name of new Databricks cluster. - :type new_cluster_name: str - :param new_cluster_node_type: The type of new Databricks cluster. - :type new_cluster_node_type: str - :param data_bricks_linked_service: Data bricks linked service. - :type data_bricks_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceResource + :param compute_type: Compute type of the cluster. The value will be overwritten by the same + setting in integration runtime if provided. + :type compute_type: str + :param core_count: Core count of the cluster. The value will be overwritten by the same setting + in integration runtime if provided. + :type core_count: int + :param time_to_live: Time to live setting of the cluster in minutes. + :type time_to_live: int + :param integration_runtime: Set to use integration runtime setting for data flow debug session. + :type integration_runtime: ~azure.synapse.artifacts.models.IntegrationRuntimeDebugResource """ _attribute_map = { - 'data_flow_name': {'key': 'dataFlowName', 'type': 'str'}, - 'existing_cluster_id': {'key': 'existingClusterId', 'type': 'str'}, - 'cluster_timeout': {'key': 'clusterTimeout', 'type': 'int'}, - 'new_cluster_name': {'key': 'newClusterName', 'type': 'str'}, - 'new_cluster_node_type': {'key': 'newClusterNodeType', 'type': 'str'}, - 'data_bricks_linked_service': {'key': 'dataBricksLinkedService', 'type': 'LinkedServiceResource'}, + 'compute_type': {'key': 'computeType', 'type': 'str'}, + 'core_count': {'key': 'coreCount', 'type': 'int'}, + 'time_to_live': {'key': 'timeToLive', 'type': 'int'}, + 'integration_runtime': {'key': 'integrationRuntime', 'type': 'IntegrationRuntimeDebugResource'}, } def __init__( self, *, - data_flow_name: Optional[str] = None, - existing_cluster_id: Optional[str] = None, - cluster_timeout: Optional[int] = None, - new_cluster_name: Optional[str] = None, - new_cluster_node_type: Optional[str] = None, - data_bricks_linked_service: Optional["LinkedServiceResource"] = None, + compute_type: Optional[str] = None, + core_count: Optional[int] = None, + time_to_live: Optional[int] = None, + integration_runtime: Optional["IntegrationRuntimeDebugResource"] = None, **kwargs ): super(CreateDataFlowDebugSessionRequest, self).__init__(**kwargs) - self.data_flow_name = data_flow_name - self.existing_cluster_id = existing_cluster_id - self.cluster_timeout = cluster_timeout - self.new_cluster_name = new_cluster_name - self.new_cluster_node_type = new_cluster_node_type - self.data_bricks_linked_service = data_bricks_linked_service + self.compute_type = compute_type + self.core_count = core_count + self.time_to_live = time_to_live + self.integration_runtime = integration_runtime class CreateDataFlowDebugSessionResponse(msrest.serialization.Model): @@ -11092,26 +11458,23 @@ class CustomActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param command: Required. Command for custom activity Type: string (or Expression with resultType string). :type command: any :param resource_linked_service: Resource linked service reference. - :type resource_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type resource_linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param folder_path: Folder path for resource files Type: string (or Expression with resultType string). :type folder_path: any :param reference_objects: Reference objects. - :type reference_objects: - ~azure.synapse.artifacts.v2021_06_01_preview.models.CustomActivityReferenceObject + :type reference_objects: ~azure.synapse.artifacts.models.CustomActivityReferenceObject :param extended_properties: User defined property bag. There is no restriction on the keys or values that can be used. The user specified custom activity has the full responsibility to consume and interpret the content defined. @@ -11182,10 +11545,9 @@ class CustomActivityReferenceObject(msrest.serialization.Model): """Reference objects for custom activity. :param linked_services: Linked service references. - :type linked_services: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference] + :type linked_services: list[~azure.synapse.artifacts.models.LinkedServiceReference] :param datasets: Dataset references. - :type datasets: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference] + :type datasets: list[~azure.synapse.artifacts.models.DatasetReference] """ _attribute_map = { @@ -11224,16 +11586,14 @@ class CustomDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param type_properties: Custom dataset properties. :type type_properties: any """ @@ -11286,13 +11646,11 @@ class CustomDataSourceLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param type_properties: Required. Custom linked service properties. @@ -11338,7 +11696,7 @@ class CustomerManagedKeyDetails(msrest.serialization.Model): :ivar status: The customer managed key status on the workspace. :vartype status: str :param key: The key object of the workspace. - :type key: ~azure.synapse.artifacts.v2021_06_01_preview.models.WorkspaceKeyDetails + :type key: ~azure.synapse.artifacts.models.WorkspaceKeyDetails """ _validation = { @@ -11377,13 +11735,11 @@ class CustomEventsTrigger(MultiplePipelineTrigger): :type description: str :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRuntimeState + :vartype runtime_state: str or ~azure.synapse.artifacts.models.TriggerRuntimeState :param annotations: List of tags that can be used for describing the trigger. :type annotations: list[any] :param pipelines: Pipelines that need to be started. - :type pipelines: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerPipelineReference] + :type pipelines: list[~azure.synapse.artifacts.models.TriggerPipelineReference] :param subject_begins_with: The event subject must begin with the pattern provided for trigger to fire. At least one of these must be provided: subjectBeginsWith, subjectEndsWith. :type subject_begins_with: str @@ -11484,14 +11840,13 @@ class DatabricksNotebookActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param notebook_path: Required. The absolute path of the notebook to be run in the Databricks Workspace. This path must begin with a slash. Type: string (or Expression with resultType string). @@ -11560,14 +11915,13 @@ class DatabricksSparkJarActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param main_class_name: Required. The full name of the class containing the main method to be executed. This class must be contained in a JAR provided as a library. Type: string (or Expression with resultType string). @@ -11635,14 +11989,13 @@ class DatabricksSparkPythonActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param python_file: Required. The URI of the Python file to be executed. DBFS paths are supported. Type: string (or Expression with resultType string). :type python_file: any @@ -11710,7 +12063,7 @@ class DataFlow(msrest.serialization.Model): :type annotations: list[any] :param folder: The folder that this data flow is in. If not specified, Data flow will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowFolder + :type folder: ~azure.synapse.artifacts.models.DataFlowFolder """ _validation = { @@ -11743,46 +12096,77 @@ def __init__( self.folder = folder -class DataFlowDebugCommandRequest(msrest.serialization.Model): - """Request body structure for data flow expression preview. +class DataFlowDebugCommandPayload(msrest.serialization.Model): + """Structure of command payload. All required parameters must be populated in order to send to Azure. - :param session_id: Required. The ID of data flow debug session. - :type session_id: str - :param data_flow_name: The data flow which contains the debug session. - :type data_flow_name: str - :param command_name: The command name. - :type command_name: str - :param command_payload: Required. The command payload object. - :type command_payload: any + :param stream_name: Required. The stream name which is used for preview. + :type stream_name: str + :param row_limits: Row limits for preview response. + :type row_limits: int + :param columns: Array of column names. + :type columns: list[str] + :param expression: The expression which is used for preview. + :type expression: str """ _validation = { - 'session_id': {'required': True}, - 'command_payload': {'required': True}, + 'stream_name': {'required': True}, } + _attribute_map = { + 'stream_name': {'key': 'streamName', 'type': 'str'}, + 'row_limits': {'key': 'rowLimits', 'type': 'int'}, + 'columns': {'key': 'columns', 'type': '[str]'}, + 'expression': {'key': 'expression', 'type': 'str'}, + } + + def __init__( + self, + *, + stream_name: str, + row_limits: Optional[int] = None, + columns: Optional[List[str]] = None, + expression: Optional[str] = None, + **kwargs + ): + super(DataFlowDebugCommandPayload, self).__init__(**kwargs) + self.stream_name = stream_name + self.row_limits = row_limits + self.columns = columns + self.expression = expression + + +class DataFlowDebugCommandRequest(msrest.serialization.Model): + """Request body structure for data flow expression preview. + + :param session_id: The ID of data flow debug session. + :type session_id: str + :param command: The command type. Possible values include: "executePreviewQuery", + "executeStatisticsQuery", "executeExpressionQuery". + :type command: str or ~azure.synapse.artifacts.models.DataFlowDebugCommandType + :param command_payload: The command payload object. + :type command_payload: ~azure.synapse.artifacts.models.DataFlowDebugCommandPayload + """ + _attribute_map = { 'session_id': {'key': 'sessionId', 'type': 'str'}, - 'data_flow_name': {'key': 'dataFlowName', 'type': 'str'}, - 'command_name': {'key': 'commandName', 'type': 'str'}, - 'command_payload': {'key': 'commandPayload', 'type': 'object'}, + 'command': {'key': 'command', 'type': 'str'}, + 'command_payload': {'key': 'commandPayload', 'type': 'DataFlowDebugCommandPayload'}, } def __init__( self, *, - session_id: str, - command_payload: Any, - data_flow_name: Optional[str] = None, - command_name: Optional[str] = None, + session_id: Optional[str] = None, + command: Optional[Union[str, "DataFlowDebugCommandType"]] = None, + command_payload: Optional["DataFlowDebugCommandPayload"] = None, **kwargs ): super(DataFlowDebugCommandRequest, self).__init__(**kwargs) self.session_id = session_id - self.data_flow_name = data_flow_name - self.command_name = command_name + self.command = command self.command_payload = command_payload @@ -11821,17 +12205,15 @@ class DataFlowDebugPackage(msrest.serialization.Model): :param session_id: The ID of data flow debug session. :type session_id: str :param data_flow: Data flow instance. - :type data_flow: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowDebugResource + :type data_flow: ~azure.synapse.artifacts.models.DataFlowDebugResource :param datasets: List of datasets. - :type datasets: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetDebugResource] + :type datasets: list[~azure.synapse.artifacts.models.DatasetDebugResource] :param linked_services: List of linked services. - :type linked_services: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceDebugResource] + :type linked_services: list[~azure.synapse.artifacts.models.LinkedServiceDebugResource] :param staging: Staging info for debug session. - :type staging: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowStagingInfo + :type staging: ~azure.synapse.artifacts.models.DataFlowStagingInfo :param debug_settings: Data flow debug settings. - :type debug_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowDebugPackageDebugSettings + :type debug_settings: ~azure.synapse.artifacts.models.DataFlowDebugPackageDebugSettings """ _attribute_map = { @@ -11870,8 +12252,7 @@ class DataFlowDebugPackageDebugSettings(msrest.serialization.Model): """Data flow debug settings. :param source_settings: Source setting for data flow debug. - :type source_settings: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowSourceSetting] + :type source_settings: list[~azure.synapse.artifacts.models.DataFlowSourceSetting] :param parameters: Data flow parameters. :type parameters: dict[str, any] :param dataset_parameters: Parameters for dataset. @@ -11984,7 +12365,7 @@ class DataFlowDebugResource(SubResourceDebugResource): :param name: The resource name. :type name: str :param properties: Required. Data flow properties. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlow + :type properties: ~azure.synapse.artifacts.models.DataFlow """ _validation = { @@ -12163,7 +12544,7 @@ class DataFlowListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of data flows. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowResource] + :type value: list[~azure.synapse.artifacts.models.DataFlowResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -12198,7 +12579,7 @@ class DataFlowReference(msrest.serialization.Model): collection. :type additional_properties: dict[str, any] :param type: Required. Data flow reference type. Possible values include: "DataFlowReference". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowReferenceType + :type type: str or ~azure.synapse.artifacts.models.DataFlowReferenceType :param reference_name: Required. Reference data flow name. :type reference_name: str :param dataset_parameters: Reference data flow parameters from dataset. @@ -12289,7 +12670,7 @@ class DataFlowResource(SubResource): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Data flow properties. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlow + :type properties: ~azure.synapse.artifacts.models.DataFlow """ _validation = { @@ -12360,13 +12741,11 @@ class DataFlowSink(Transformation): :param description: Transformation description. :type description: str :param dataset: Dataset reference. - :type dataset: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference + :type dataset: ~azure.synapse.artifacts.models.DatasetReference :param linked_service: Linked service reference. - :type linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param schema_linked_service: Schema linked service reference. - :type schema_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type schema_linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference """ _validation = { @@ -12407,13 +12786,11 @@ class DataFlowSource(Transformation): :param description: Transformation description. :type description: str :param dataset: Dataset reference. - :type dataset: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference + :type dataset: ~azure.synapse.artifacts.models.DatasetReference :param linked_service: Linked service reference. - :type linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param schema_linked_service: Schema linked service reference. - :type schema_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type schema_linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference """ _validation = { @@ -12480,8 +12857,7 @@ class DataFlowStagingInfo(msrest.serialization.Model): """Staging info for execute data flow activity. :param linked_service: Staging linked service reference. - :type linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param folder_path: Folder path for staging blob. :type folder_path: str """ @@ -12518,20 +12894,18 @@ class DataLakeAnalyticsUSQLActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param script_path: Required. Case-sensitive path to folder that contains the U-SQL script. Type: string (or Expression with resultType string). :type script_path: any :param script_linked_service: Required. Script linked service reference. - :type script_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type script_linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param degree_of_parallelism: The maximum number of nodes simultaneously used to run the job. Default value is 1. Type: integer (or Expression with resultType integer), minimum: 1. :type degree_of_parallelism: any @@ -12641,9 +13015,8 @@ class DatasetCompression(msrest.serialization.Model): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any + :param type: Required. Type of dataset compression.Constant filled by server. + :type type: str """ _validation = { @@ -12652,7 +13025,7 @@ class DatasetCompression(msrest.serialization.Model): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, } _subtype_map = { @@ -12678,9 +13051,8 @@ class DatasetBZip2Compression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any + :param type: Required. Type of dataset compression.Constant filled by server. + :type type: str """ _validation = { @@ -12689,7 +13061,7 @@ class DatasetBZip2Compression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, } def __init__( @@ -12736,7 +13108,7 @@ class DatasetDebugResource(SubResourceDebugResource): :param name: The resource name. :type name: str :param properties: Required. Dataset properties. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.Dataset + :type properties: ~azure.synapse.artifacts.models.Dataset """ _validation = { @@ -12767,10 +13139,10 @@ class DatasetDeflateCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - :param level: The Deflate compression level. + :param type: Required. Type of dataset compression.Constant filled by server. + :type type: str + :param level: The Deflate compression level. Type: string (or Expression with resultType + string). :type level: any """ @@ -12780,7 +13152,7 @@ class DatasetDeflateCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, 'level': {'key': 'level', 'type': 'object'}, } @@ -12825,10 +13197,9 @@ class DatasetGZipCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - :param level: The GZip compression level. + :param type: Required. Type of dataset compression.Constant filled by server. + :type type: str + :param level: The GZip compression level. Type: string (or Expression with resultType string). :type level: any """ @@ -12838,7 +13209,7 @@ class DatasetGZipCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, 'level': {'key': 'level', 'type': 'object'}, } @@ -12860,7 +13231,7 @@ class DatasetListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of datasets. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetResource] + :type value: list[~azure.synapse.artifacts.models.DatasetResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -12892,7 +13263,7 @@ class DatasetReference(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param type: Required. Dataset reference type. Possible values include: "DatasetReference". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReferenceType + :type type: str or ~azure.synapse.artifacts.models.DatasetReferenceType :param reference_name: Required. Reference dataset name. :type reference_name: str :param parameters: Arguments for dataset. @@ -12942,7 +13313,7 @@ class DatasetResource(SubResource): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Dataset properties. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.Dataset + :type properties: ~azure.synapse.artifacts.models.Dataset """ _validation = { @@ -13011,9 +13382,8 @@ class DatasetTarCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any + :param type: Required. Type of dataset compression.Constant filled by server. + :type type: str """ _validation = { @@ -13022,7 +13392,7 @@ class DatasetTarCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, } def __init__( @@ -13043,10 +13413,10 @@ class DatasetTarGZipCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - :param level: The TarGZip compression level. + :param type: Required. Type of dataset compression.Constant filled by server. + :type type: str + :param level: The TarGZip compression level. Type: string (or Expression with resultType + string). :type level: any """ @@ -13056,7 +13426,7 @@ class DatasetTarGZipCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, 'level': {'key': 'level', 'type': 'object'}, } @@ -13080,10 +13450,10 @@ class DatasetZipDeflateCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - :param level: The ZipDeflate compression level. + :param type: Required. Type of dataset compression.Constant filled by server. + :type type: str + :param level: The ZipDeflate compression level. Type: string (or Expression with resultType + string). :type level: any """ @@ -13093,7 +13463,7 @@ class DatasetZipDeflateCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, 'level': {'key': 'level', 'type': 'object'}, } @@ -13120,13 +13490,11 @@ class Db2LinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: The connection string. It is mutually exclusive with server, @@ -13141,13 +13509,12 @@ class Db2LinkedService(LinkedService): :type database: any :param authentication_type: AuthenticationType to be used for connection. It is mutually exclusive with connectionString property. Possible values include: "Basic". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.Db2AuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.Db2AuthenticationType :param username: Username for authentication. It is mutually exclusive with connectionString property. Type: string (or Expression with resultType string). :type username: any :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param package_collection: Under where packages are created when querying database. It is mutually exclusive with connectionString property. Type: string (or Expression with resultType string). @@ -13297,16 +13664,14 @@ class Db2TableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -13375,14 +13740,13 @@ class DeleteActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param recursive: If true, files or sub-folders under current folder path will be deleted recursively. Default is false. Type: boolean (or Expression with resultType boolean). :type recursive: any @@ -13394,12 +13758,11 @@ class DeleteActivity(ExecutionActivity): :type enable_logging: any :param log_storage_settings: Log storage settings customer need to provide when enableLogging is true. - :type log_storage_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LogStorageSettings + :type log_storage_settings: ~azure.synapse.artifacts.models.LogStorageSettings :param dataset: Required. Delete activity dataset reference. - :type dataset: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference + :type dataset: ~azure.synapse.artifacts.models.DatasetReference :param store_settings: Delete activity store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings """ _validation = { @@ -13499,18 +13862,16 @@ class DelimitedTextDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param location: The location of the delimited text storage. - :type location: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetLocation + :type location: ~azure.synapse.artifacts.models.DatasetLocation :param column_delimiter: The column delimiter. Type: string (or Expression with resultType string). :type column_delimiter: any @@ -13522,10 +13883,9 @@ class DelimitedTextDataset(Dataset): https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string (or Expression with resultType string). :type encoding_name: any - :param compression_codec: Possible values include: "bzip2", "gzip", "deflate", "zipDeflate", - "snappy", "lz4", "tar", "tarGZip". - :type compression_codec: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.CompressionCodec + :param compression_codec: The data compressionCodec. Type: string (or Expression with + resultType string). + :type compression_codec: any :param compression_level: The data compression method used for DelimitedText. :type compression_level: any :param quote_char: The quote character. Type: string (or Expression with resultType string). @@ -13559,7 +13919,7 @@ class DelimitedTextDataset(Dataset): 'column_delimiter': {'key': 'typeProperties.columnDelimiter', 'type': 'object'}, 'row_delimiter': {'key': 'typeProperties.rowDelimiter', 'type': 'object'}, 'encoding_name': {'key': 'typeProperties.encodingName', 'type': 'object'}, - 'compression_codec': {'key': 'typeProperties.compressionCodec', 'type': 'str'}, + 'compression_codec': {'key': 'typeProperties.compressionCodec', 'type': 'object'}, 'compression_level': {'key': 'typeProperties.compressionLevel', 'type': 'object'}, 'quote_char': {'key': 'typeProperties.quoteChar', 'type': 'object'}, 'escape_char': {'key': 'typeProperties.escapeChar', 'type': 'object'}, @@ -13582,7 +13942,7 @@ def __init__( column_delimiter: Optional[Any] = None, row_delimiter: Optional[Any] = None, encoding_name: Optional[Any] = None, - compression_codec: Optional[Union[str, "CompressionCodec"]] = None, + compression_codec: Optional[Any] = None, compression_level: Optional[Any] = None, quote_char: Optional[Any] = None, escape_char: Optional[Any] = None, @@ -13618,8 +13978,7 @@ class DelimitedTextReadSettings(FormatReadSettings): input files. Type: integer (or Expression with resultType integer). :type skip_line_count: any :param compression_properties: Compression settings. - :type compression_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.CompressionReadSettings + :type compression_properties: ~azure.synapse.artifacts.models.CompressionReadSettings """ _validation = { @@ -13673,10 +14032,9 @@ class DelimitedTextSink(CopySink): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: DelimitedText store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreWriteSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreWriteSettings :param format_settings: DelimitedText format settings. - :type format_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.DelimitedTextWriteSettings + :type format_settings: ~azure.synapse.artifacts.models.DelimitedTextWriteSettings """ _validation = { @@ -13734,10 +14092,9 @@ class DelimitedTextSource(CopySource): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: DelimitedText store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings :param format_settings: DelimitedText format settings. - :type format_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.DelimitedTextReadSettings + :type format_settings: ~azure.synapse.artifacts.models.DelimitedTextReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :type additional_columns: any @@ -13927,16 +14284,14 @@ class DocumentDbCollectionDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param collection_name: Required. Document Database collection name. Type: string (or Expression with resultType string). :type collection_name: any @@ -14128,20 +14483,18 @@ class DrillLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type pwd: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -14265,16 +14618,14 @@ class DrillTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -14362,8 +14713,7 @@ class DWCopyCommandSettings(msrest.serialization.Model): default values in the property overwrite the DEFAULT constraint set in the DB, and identity column cannot have a default value. Type: array of objects (or Expression with resultType array of objects). - :type default_values: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.DWCopyCommandDefaultValue] + :type default_values: list[~azure.synapse.artifacts.models.DWCopyCommandDefaultValue] :param additional_options: Additional options directly passed to SQL DW in Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalOptions": { "MAXERRORS": "1000", "DATEFORMAT": "'ymd'" }. @@ -14419,13 +14769,11 @@ class DynamicsAXLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param url: Required. The Dynamics AX (or Dynamics 365 Finance and Operations) instance OData @@ -14437,7 +14785,7 @@ class DynamicsAXLinkedService(LinkedService): :param service_principal_key: Required. Specify the application's key. Mark this field as a SecureString to store it securely in Data Factory, or reference a secret stored in Azure Key Vault. Type: string (or Expression with resultType string). - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: Required. Specify the tenant information (domain name or tenant ID) under which your application resides. Retrieve it by hovering the mouse in the top-right corner of the Azure portal. Type: string (or Expression with resultType string). @@ -14520,16 +14868,14 @@ class DynamicsAXResourceDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param path: Required. The path of the Dynamics AX OData entity. Type: string (or Expression with resultType string). :type path: any @@ -14662,16 +15008,14 @@ class DynamicsCrmEntityDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param entity_name: The logical name of the entity. Type: string (or Expression with resultType string). :type entity_name: any @@ -14725,21 +15069,17 @@ class DynamicsCrmLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param deployment_type: Required. The deployment type of the Dynamics CRM instance. 'Online' for Dynamics CRM Online and 'OnPremisesWithIfd' for Dynamics CRM on-premises with Ifd. Type: - string (or Expression with resultType string). Possible values include: "Online", - "OnPremisesWithIfd". - :type deployment_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsDeploymentType + string (or Expression with resultType string). + :type deployment_type: any :param host_name: The host name of the on-premises Dynamics CRM server. The property is required for on-prem and not allowed for online. Type: string (or Expression with resultType string). @@ -14758,15 +15098,13 @@ class DynamicsCrmLinkedService(LinkedService): :param authentication_type: Required. The authentication type to connect to Dynamics CRM server. 'Office365' for online scenario, 'Ifd' for on-premises with Ifd scenario, 'AADServicePrincipal' for Server-To-Server authentication in online scenario. Type: string (or - Expression with resultType string). Possible values include: "Office365", "Ifd", - "AADServicePrincipal". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsAuthenticationType + Expression with resultType string). + :type authentication_type: any :param username: User name to access the Dynamics CRM instance. Type: string (or Expression with resultType string). :type username: any :param password: Password to access the Dynamics CRM instance. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param service_principal_id: The client ID of the application in Azure Active Directory used for Server-To-Server authentication. Type: string (or Expression with resultType string). :type service_principal_id: any @@ -14778,8 +15116,7 @@ class DynamicsCrmLinkedService(LinkedService): servicePrincipalCredential can be SecureString or AzureKeyVaultSecretReference. If servicePrincipalCredentialType is 'ServicePrincipalCert', servicePrincipalCredential can only be AzureKeyVaultSecretReference. - :type service_principal_credential: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_credential: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -14799,12 +15136,12 @@ class DynamicsCrmLinkedService(LinkedService): 'description': {'key': 'description', 'type': 'str'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'str'}, + 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'object'}, 'host_name': {'key': 'typeProperties.hostName', 'type': 'object'}, 'port': {'key': 'typeProperties.port', 'type': 'object'}, 'service_uri': {'key': 'typeProperties.serviceUri', 'type': 'object'}, 'organization_name': {'key': 'typeProperties.organizationName', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, 'username': {'key': 'typeProperties.username', 'type': 'object'}, 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, @@ -14816,8 +15153,8 @@ class DynamicsCrmLinkedService(LinkedService): def __init__( self, *, - deployment_type: Union[str, "DynamicsDeploymentType"], - authentication_type: Union[str, "DynamicsAuthenticationType"], + deployment_type: Any, + authentication_type: Any, additional_properties: Optional[Dict[str, Any]] = None, connect_via: Optional["IntegrationRuntimeReference"] = None, description: Optional[str] = None, @@ -14878,8 +15215,7 @@ class DynamicsCrmSink(CopySink): :type max_concurrent_connections: any :param write_behavior: Required. The write behavior for the operation. Possible values include: "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsSinkWriteBehavior + :type write_behavior: str or ~azure.synapse.artifacts.models.DynamicsSinkWriteBehavior :param ignore_null_values: The flag indicating whether to ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). @@ -15005,16 +15341,14 @@ class DynamicsEntityDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param entity_name: The logical name of the entity. Type: string (or Expression with resultType string). :type entity_name: any @@ -15068,20 +15402,17 @@ class DynamicsLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param deployment_type: Required. The deployment type of the Dynamics instance. 'Online' for Dynamics Online and 'OnPremisesWithIfd' for Dynamics on-premises with Ifd. Type: string (or - Expression with resultType string). Possible values include: "Online", "OnPremisesWithIfd". - :type deployment_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsDeploymentType + Expression with resultType string). + :type deployment_type: any :param host_name: The host name of the on-premises Dynamics server. The property is required for on-prem and not allowed for online. Type: string (or Expression with resultType string). :type host_name: any @@ -15099,30 +15430,26 @@ class DynamicsLinkedService(LinkedService): :param authentication_type: Required. The authentication type to connect to Dynamics server. 'Office365' for online scenario, 'Ifd' for on-premises with Ifd scenario, 'AADServicePrincipal' for Server-To-Server authentication in online scenario. Type: string (or Expression with - resultType string). Possible values include: "Office365", "Ifd", "AADServicePrincipal". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsAuthenticationType + resultType string). + :type authentication_type: any :param username: User name to access the Dynamics instance. Type: string (or Expression with resultType string). :type username: any :param password: Password to access the Dynamics instance. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param service_principal_id: The client ID of the application in Azure Active Directory used for Server-To-Server authentication. Type: string (or Expression with resultType string). :type service_principal_id: any :param service_principal_credential_type: The service principal credential type to use in Server-To-Server authentication. 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' - for certificate. Type: string (or Expression with resultType string). Possible values include: - "ServicePrincipalKey", "ServicePrincipalCert". - :type service_principal_credential_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsServicePrincipalCredentialType + for certificate. Type: string (or Expression with resultType string). + :type service_principal_credential_type: any :param service_principal_credential: The credential of the service principal object in Azure Active Directory. If servicePrincipalCredentialType is 'ServicePrincipalKey', servicePrincipalCredential can be SecureString or AzureKeyVaultSecretReference. If servicePrincipalCredentialType is 'ServicePrincipalCert', servicePrincipalCredential can only be AzureKeyVaultSecretReference. - :type service_principal_credential: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_credential: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -15142,16 +15469,16 @@ class DynamicsLinkedService(LinkedService): 'description': {'key': 'description', 'type': 'str'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'str'}, + 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'object'}, 'host_name': {'key': 'typeProperties.hostName', 'type': 'object'}, 'port': {'key': 'typeProperties.port', 'type': 'object'}, 'service_uri': {'key': 'typeProperties.serviceUri', 'type': 'object'}, 'organization_name': {'key': 'typeProperties.organizationName', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, 'username': {'key': 'typeProperties.username', 'type': 'object'}, 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_credential_type': {'key': 'typeProperties.servicePrincipalCredentialType', 'type': 'str'}, + 'service_principal_credential_type': {'key': 'typeProperties.servicePrincipalCredentialType', 'type': 'object'}, 'service_principal_credential': {'key': 'typeProperties.servicePrincipalCredential', 'type': 'SecretBase'}, 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } @@ -15159,8 +15486,8 @@ class DynamicsLinkedService(LinkedService): def __init__( self, *, - deployment_type: Union[str, "DynamicsDeploymentType"], - authentication_type: Union[str, "DynamicsAuthenticationType"], + deployment_type: Any, + authentication_type: Any, additional_properties: Optional[Dict[str, Any]] = None, connect_via: Optional["IntegrationRuntimeReference"] = None, description: Optional[str] = None, @@ -15173,7 +15500,7 @@ def __init__( username: Optional[Any] = None, password: Optional["SecretBase"] = None, service_principal_id: Optional[Any] = None, - service_principal_credential_type: Optional[Union[str, "DynamicsServicePrincipalCredentialType"]] = None, + service_principal_credential_type: Optional[Any] = None, service_principal_credential: Optional["SecretBase"] = None, encrypted_credential: Optional[Any] = None, **kwargs @@ -15221,8 +15548,7 @@ class DynamicsSink(CopySink): :type max_concurrent_connections: any :param write_behavior: Required. The write behavior for the operation. Possible values include: "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DynamicsSinkWriteBehavior + :type write_behavior: str or ~azure.synapse.artifacts.models.DynamicsSinkWriteBehavior :param ignore_null_values: The flag indicating whether ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). @@ -15340,13 +15666,11 @@ class EloquaLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param endpoint: Required. The endpoint of the Eloqua server. (i.e. eloqua.example.com). @@ -15355,7 +15679,7 @@ class EloquaLinkedService(LinkedService): sitename/username. (i.e. Eloqua/Alice). :type username: any :param password: The password corresponding to the user name. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -15441,16 +15765,14 @@ class EloquaObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -15562,7 +15884,7 @@ class EncryptionDetails(msrest.serialization.Model): :ivar double_encryption_enabled: Double Encryption enabled. :vartype double_encryption_enabled: bool :param cmk: Customer Managed Key Details. - :type cmk: ~azure.synapse.artifacts.v2021_06_01_preview.models.CustomerManagedKeyDetails + :type cmk: ~azure.synapse.artifacts.models.CustomerManagedKeyDetails """ _validation = { @@ -15590,8 +15912,7 @@ class EntityReference(msrest.serialization.Model): :param type: The type of this referenced entity. Possible values include: "IntegrationRuntimeReference", "LinkedServiceReference". - :type type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeEntityReferenceType + :type type: str or ~azure.synapse.artifacts.models.IntegrationRuntimeEntityReferenceType :param reference_name: The name of this referenced entity. :type reference_name: str """ @@ -15647,7 +15968,7 @@ class ErrorContract(msrest.serialization.Model): """Contains details when the response code indicates an error. :param error: The error details. - :type error: ~azure.synapse.artifacts.v2021_06_01_preview.models.ErrorResponse + :type error: ~azure.synapse.artifacts.models.ErrorResponse """ _attribute_map = { @@ -15676,10 +15997,9 @@ class ErrorResponse(msrest.serialization.Model): :ivar target: The error target. :vartype target: str :ivar details: The error details. - :vartype details: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ErrorResponse] + :vartype details: list[~azure.synapse.artifacts.models.ErrorResponse] :ivar additional_info: The error additional info. - :vartype additional_info: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.ErrorAdditionalInfo] + :vartype additional_info: list[~azure.synapse.artifacts.models.ErrorAdditionalInfo] """ _validation = { @@ -15770,18 +16090,16 @@ class ExcelDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param location: The location of the excel storage. - :type location: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetLocation + :type location: ~azure.synapse.artifacts.models.DatasetLocation :param sheet_name: The sheet of excel file. Type: string (or Expression with resultType string). :type sheet_name: any @@ -15793,7 +16111,7 @@ class ExcelDataset(Dataset): false. Type: boolean (or Expression with resultType boolean). :type first_row_as_header: any :param compression: The data compression method used for the json dataset. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression :param null_value: The null value string. Type: string (or Expression with resultType string). :type null_value: any """ @@ -15870,7 +16188,7 @@ class ExcelSource(CopySource): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Excel store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :type additional_columns: any @@ -15922,24 +16240,21 @@ class ExecuteDataFlowActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param dataflow: Required. Data flow reference. - :type dataflow: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowReference + :type dataflow: ~azure.synapse.artifacts.models.DataFlowReference :param staging: Staging info for execute data flow activity. - :type staging: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowStagingInfo + :type staging: ~azure.synapse.artifacts.models.DataFlowStagingInfo :param integration_runtime: The integration runtime reference. - :type integration_runtime: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type integration_runtime: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param compute: Compute properties for data flow activity. - :type compute: - ~azure.synapse.artifacts.v2021_06_01_preview.models.ExecuteDataFlowActivityTypePropertiesCompute + :type compute: ~azure.synapse.artifacts.models.ExecuteDataFlowActivityTypePropertiesCompute :param trace_level: Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string). :type trace_level: any @@ -16011,8 +16326,7 @@ class ExecuteDataFlowActivityTypePropertiesCompute(msrest.serialization.Model): :param compute_type: Compute type of the cluster which will execute data flow job. Possible values include: "General", "MemoryOptimized", "ComputeOptimized". - :type compute_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowComputeType + :type compute_type: str or ~azure.synapse.artifacts.models.DataFlowComputeType :param core_count: Core count of the cluster which will execute data flow job. Supported values are: 8, 16, 32, 48, 80, 144 and 272. :type core_count: int @@ -16050,11 +16364,11 @@ class ExecutePipelineActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param pipeline: Required. Pipeline reference. - :type pipeline: ~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineReference + :type pipeline: ~azure.synapse.artifacts.models.PipelineReference :param parameters: Pipeline parameters. :type parameters: dict[str, any] :param wait_on_completion: Defines whether activity execution will wait for the dependent @@ -16115,16 +16429,15 @@ class ExecuteSSISPackageActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param package_location: Required. SSIS package location. - :type package_location: ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISPackageLocation + :type package_location: ~azure.synapse.artifacts.models.SSISPackageLocation :param runtime: Specifies the runtime to execute SSIS package. The value should be "x86" or "x64". Type: string (or Expression with resultType string). :type runtime: any @@ -16135,30 +16448,25 @@ class ExecuteSSISPackageActivity(ExecutionActivity): Expression with resultType string). :type environment_path: any :param execution_credential: The package execution credential. - :type execution_credential: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISExecutionCredential + :type execution_credential: ~azure.synapse.artifacts.models.SSISExecutionCredential :param connect_via: Required. The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param project_parameters: The project level parameters to execute the SSIS package. - :type project_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISExecutionParameter] + :type project_parameters: dict[str, ~azure.synapse.artifacts.models.SSISExecutionParameter] :param package_parameters: The package level parameters to execute the SSIS package. - :type package_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISExecutionParameter] + :type package_parameters: dict[str, ~azure.synapse.artifacts.models.SSISExecutionParameter] :param project_connection_managers: The project level connection managers to execute the SSIS package. :type project_connection_managers: dict[str, dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISExecutionParameter]] + ~azure.synapse.artifacts.models.SSISExecutionParameter]] :param package_connection_managers: The package level connection managers to execute the SSIS package. :type package_connection_managers: dict[str, dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISExecutionParameter]] + ~azure.synapse.artifacts.models.SSISExecutionParameter]] :param property_overrides: The property overrides to execute the SSIS package. - :type property_overrides: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISPropertyOverride] + :type property_overrides: dict[str, ~azure.synapse.artifacts.models.SSISPropertyOverride] :param log_location: SSIS package execution log location. - :type log_location: ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISLogLocation + :type log_location: ~azure.synapse.artifacts.models.SSISLogLocation """ _validation = { @@ -16293,7 +16601,7 @@ class Expression(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param type: Required. Expression type. Possible values include: "Expression". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.ExpressionType + :type type: str or ~azure.synapse.artifacts.models.ExpressionType :param value: Required. Expression value. :type value: str """ @@ -16331,13 +16639,11 @@ class FileServerLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. Host name of the server. Type: string (or Expression with resultType @@ -16347,7 +16653,7 @@ class FileServerLinkedService(LinkedService): string). :type user_id: any :param password: Password to logon the server. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -16590,16 +16896,14 @@ class FileShareDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param folder_path: The path of the on-premises file system. Type: string (or Expression with resultType string). :type folder_path: any @@ -16613,12 +16917,12 @@ class FileShareDataset(Dataset): with resultType string). :type modified_datetime_end: any :param format: The format of the files. - :type format: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetStorageFormat + :type format: ~azure.synapse.artifacts.models.DatasetStorageFormat :param file_filter: Specify a filter to be used to select a subset of files in the folderPath rather than all files. Type: string (or Expression with resultType string). :type file_filter: any :param compression: The data compression method used for the file system. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression """ _validation = { @@ -16810,13 +17114,13 @@ class FilterActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param items: Required. Input array on which filter should be applied. - :type items: ~azure.synapse.artifacts.v2021_06_01_preview.models.Expression + :type items: ~azure.synapse.artifacts.models.Expression :param condition: Required. Condition to be used for filtering the input. - :type condition: ~azure.synapse.artifacts.v2021_06_01_preview.models.Expression + :type condition: ~azure.synapse.artifacts.models.Expression """ _validation = { @@ -16870,18 +17174,18 @@ class ForEachActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param is_sequential: Should the loop be executed in sequence or in parallel (max 50). :type is_sequential: bool :param batch_count: Batch count to be used for controlling the number of parallel execution (when isSequential is set to false). :type batch_count: int :param items: Required. Collection to iterate. - :type items: ~azure.synapse.artifacts.v2021_06_01_preview.models.Expression + :type items: ~azure.synapse.artifacts.models.Expression :param activities: Required. List of activities to execute . - :type activities: list[~azure.synapse.artifacts.v2021_06_01_preview.models.Activity] + :type activities: list[~azure.synapse.artifacts.models.Activity] """ _validation = { @@ -17021,13 +17325,11 @@ class FtpServerLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. Host name of the FTP server. Type: string (or Expression with resultType @@ -17038,13 +17340,12 @@ class FtpServerLinkedService(LinkedService): :type port: any :param authentication_type: The authentication type to be used to connect to the FTP server. Possible values include: "Basic", "Anonymous". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.FtpAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.FtpAuthenticationType :param user_name: Username to logon the FTP server. Type: string (or Expression with resultType string). :type user_name: any :param password: Password to logon the FTP server. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -17166,22 +17467,21 @@ class GetMetadataActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param dataset: Required. GetMetadata activity dataset reference. - :type dataset: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference + :type dataset: ~azure.synapse.artifacts.models.DatasetReference :param field_list: Fields of metadata to get from dataset. :type field_list: list[any] :param store_settings: GetMetadata activity store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings :param format_settings: GetMetadata activity format settings. - :type format_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.FormatReadSettings + :type format_settings: ~azure.synapse.artifacts.models.FormatReadSettings """ _validation = { @@ -17347,13 +17647,11 @@ class GoogleAdWordsLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param client_customer_id: Required. The Client customer ID of the AdWords account that you @@ -17361,21 +17659,21 @@ class GoogleAdWordsLinkedService(LinkedService): :type client_customer_id: any :param developer_token: Required. The developer token associated with the manager account that you use to grant access to the AdWords API. - :type developer_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type developer_token: ~azure.synapse.artifacts.models.SecretBase :param authentication_type: Required. The OAuth 2.0 authentication mechanism used for authentication. ServiceAuthentication can only be used on self-hosted IR. Possible values include: "ServiceAuthentication", "UserAuthentication". :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.GoogleAdWordsAuthenticationType + ~azure.synapse.artifacts.models.GoogleAdWordsAuthenticationType :param refresh_token: The refresh token obtained from Google for authorizing access to AdWords for UserAuthentication. - :type refresh_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type refresh_token: ~azure.synapse.artifacts.models.SecretBase :param client_id: The client id of the google application used to acquire the refresh token. Type: string (or Expression with resultType string). :type client_id: any :param client_secret: The client secret of the google application used to acquire the refresh token. - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type client_secret: ~azure.synapse.artifacts.models.SecretBase :param email: The service account email ID that is used for ServiceAuthentication and can only be used on self-hosted IR. :type email: any @@ -17477,16 +17775,14 @@ class GoogleAdWordsObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -17601,13 +17897,11 @@ class GoogleBigQueryLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param project: Required. The default BigQuery project to query against. @@ -17622,16 +17916,16 @@ class GoogleBigQueryLinkedService(LinkedService): authentication. ServiceAuthentication can only be used on self-hosted IR. Possible values include: "ServiceAuthentication", "UserAuthentication". :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.GoogleBigQueryAuthenticationType + ~azure.synapse.artifacts.models.GoogleBigQueryAuthenticationType :param refresh_token: The refresh token obtained from Google for authorizing access to BigQuery for UserAuthentication. - :type refresh_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type refresh_token: ~azure.synapse.artifacts.models.SecretBase :param client_id: The client id of the google application used to acquire the refresh token. Type: string (or Expression with resultType string). :type client_id: any :param client_secret: The client secret of the google application used to acquire the refresh token. - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type client_secret: ~azure.synapse.artifacts.models.SecretBase :param email: The service account email ID that is used for ServiceAuthentication and can only be used on self-hosted IR. :type email: any @@ -17735,16 +18029,14 @@ class GoogleBigQueryObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using database + table properties instead. :type table_name: any @@ -17872,13 +18164,11 @@ class GoogleCloudStorageLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param access_key_id: The access key identifier of the Google Cloud Storage Identity and Access @@ -17886,7 +18176,7 @@ class GoogleCloudStorageLinkedService(LinkedService): :type access_key_id: any :param secret_access_key: The secret access key of the Google Cloud Storage Identity and Access Management (IAM) user. - :type secret_access_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type secret_access_key: ~azure.synapse.artifacts.models.SecretBase :param service_url: This value specifies the endpoint to access with the Google Cloud Storage Connector. This is an optional property; change it only if you want to try a different service endpoint or want to switch between https and http. Type: string (or Expression with resultType @@ -18097,20 +18387,18 @@ class GreenplumLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type pwd: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -18234,16 +18522,14 @@ class GreenplumTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -18308,13 +18594,11 @@ class HBaseLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The IP address or host name of the HBase server. (i.e. 192.168.222.160). @@ -18327,12 +18611,11 @@ class HBaseLinkedService(LinkedService): :type http_path: any :param authentication_type: Required. The authentication mechanism to use to connect to the HBase server. Possible values include: "Anonymous", "Basic". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.HBaseAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.HBaseAuthenticationType :param username: The user name used to connect to the HBase instance. :type username: any :param password: The password corresponding to the user name. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The default value is false. :type enable_ssl: any @@ -18433,16 +18716,14 @@ class HBaseObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -18557,13 +18838,11 @@ class HdfsLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param url: Required. The URL of the HDFS service endpoint, e.g. @@ -18580,7 +18859,7 @@ class HdfsLinkedService(LinkedService): resultType string). :type user_name: any :param password: Password for Windows authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase """ _validation = { @@ -18705,7 +18984,7 @@ class HdfsReadSettings(StoreReadSettings): with resultType string). :type modified_datetime_end: any :param distcp_settings: Specifies Distcp-related settings. - :type distcp_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.DistcpSettings + :type distcp_settings: ~azure.synapse.artifacts.models.DistcpSettings :param delete_files_after_completion: Indicates whether the source files need to be deleted after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). :type delete_files_after_completion: any @@ -18785,7 +19064,7 @@ class HdfsSource(CopySource): true. Type: boolean (or Expression with resultType boolean). :type recursive: any :param distcp_settings: Specifies Distcp-related settings. - :type distcp_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.DistcpSettings + :type distcp_settings: ~azure.synapse.artifacts.models.DistcpSettings """ _validation = { @@ -18834,27 +19113,23 @@ class HDInsightHiveActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param storage_linked_services: Storage linked service references. - :type storage_linked_services: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference] + :type storage_linked_services: list[~azure.synapse.artifacts.models.LinkedServiceReference] :param arguments: User specified arguments to HDInsightActivity. :type arguments: list[any] :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.HDInsightActivityDebugInfoOption + :type get_debug_info: str or ~azure.synapse.artifacts.models.HDInsightActivityDebugInfoOption :param script_path: Script path. Type: string (or Expression with resultType string). :type script_path: any :param script_linked_service: Script linked service reference. - :type script_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type script_linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param defines: Allows user to specify defines for Hive job request. :type defines: dict[str, any] :param variables: User specified arguments under hivevar namespace. @@ -18931,13 +19206,11 @@ class HDInsightLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param cluster_uri: Required. HDInsight cluster URI. Type: string (or Expression with @@ -18947,14 +19220,12 @@ class HDInsightLinkedService(LinkedService): string). :type user_name: any :param password: HDInsight cluster password. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param linked_service_name: The Azure Storage linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param hcatalog_linked_service_name: A reference to the Azure SQL linked service that points to the HCatalog database. - :type hcatalog_linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type hcatalog_linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -19034,29 +19305,25 @@ class HDInsightMapReduceActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param storage_linked_services: Storage linked service references. - :type storage_linked_services: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference] + :type storage_linked_services: list[~azure.synapse.artifacts.models.LinkedServiceReference] :param arguments: User specified arguments to HDInsightActivity. :type arguments: list[any] :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.HDInsightActivityDebugInfoOption + :type get_debug_info: str or ~azure.synapse.artifacts.models.HDInsightActivityDebugInfoOption :param class_name: Required. Class name. Type: string (or Expression with resultType string). :type class_name: any :param jar_file_path: Required. Jar path. Type: string (or Expression with resultType string). :type jar_file_path: any :param jar_linked_service: Jar linked service reference. - :type jar_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type jar_linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param jar_libs: Jar libs. :type jar_libs: list[any] :param defines: Allows user to specify defines for the MapReduce job request. @@ -19132,13 +19399,11 @@ class HDInsightOnDemandLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param cluster_size: Required. Number of worker/data nodes in the cluster. Suggestion value: 4. @@ -19154,8 +19419,7 @@ class HDInsightOnDemandLinkedService(LinkedService): :type version: any :param linked_service_name: Required. Azure Storage linked service to be used by the on-demand cluster for storing and processing data. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param host_subscription_id: Required. The customer’s subscription to host the cluster. Type: string (or Expression with resultType string). :type host_subscription_id: any @@ -19163,7 +19427,7 @@ class HDInsightOnDemandLinkedService(LinkedService): (or Expression with resultType string). :type service_principal_id: any :param service_principal_key: The key for the service principal id. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: Required. The Tenant id/name to which the service principal belongs. Type: string (or Expression with resultType string). :type tenant: any @@ -19177,21 +19441,20 @@ class HDInsightOnDemandLinkedService(LinkedService): resultType string). :type cluster_user_name: any :param cluster_password: The password to access the cluster. - :type cluster_password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type cluster_password: ~azure.synapse.artifacts.models.SecretBase :param cluster_ssh_user_name: The username to SSH remotely connect to cluster’s node (for Linux). Type: string (or Expression with resultType string). :type cluster_ssh_user_name: any :param cluster_ssh_password: The password to SSH remotely connect cluster’s node (for Linux). - :type cluster_ssh_password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type cluster_ssh_password: ~azure.synapse.artifacts.models.SecretBase :param additional_linked_service_names: Specifies additional storage accounts for the HDInsight linked service so that the Data Factory service can register them on your behalf. :type additional_linked_service_names: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference] + list[~azure.synapse.artifacts.models.LinkedServiceReference] :param hcatalog_linked_service_name: The name of Azure SQL linked service that point to the HCatalog database. The on-demand HDInsight cluster is created by using the Azure SQL database as the metastore. - :type hcatalog_linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type hcatalog_linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param cluster_type: The cluster type. Type: string (or Expression with resultType string). :type cluster_type: any :param spark_version: The version of spark if the cluster type is 'spark'. Type: string (or @@ -19235,7 +19498,7 @@ class HDInsightOnDemandLinkedService(LinkedService): :param script_actions: Custom script actions to run on HDI ondemand cluster once it's up. Please refer to https://docs.microsoft.com/en-us/azure/hdinsight/hdinsight-hadoop-customize-cluster-linux?toc=%2Fen-us%2Fazure%2Fhdinsight%2Fr-server%2FTOC.json&bc=%2Fen-us%2Fazure%2Fbread%2Ftoc.json#understanding-script-actions. - :type script_actions: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ScriptAction] + :type script_actions: list[~azure.synapse.artifacts.models.ScriptAction] :param virtual_network_id: The ARM resource ID for the vNet to which the cluster should be joined after creation. Type: string (or Expression with resultType string). :type virtual_network_id: any @@ -19392,28 +19655,24 @@ class HDInsightPigActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param storage_linked_services: Storage linked service references. - :type storage_linked_services: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference] + :type storage_linked_services: list[~azure.synapse.artifacts.models.LinkedServiceReference] :param arguments: User specified arguments to HDInsightActivity. Type: array (or Expression with resultType array). :type arguments: any :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.HDInsightActivityDebugInfoOption + :type get_debug_info: str or ~azure.synapse.artifacts.models.HDInsightActivityDebugInfoOption :param script_path: Script path. Type: string (or Expression with resultType string). :type script_path: any :param script_linked_service: Script linked service reference. - :type script_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type script_linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param defines: Allows user to specify defines for Pig job request. :type defines: dict[str, any] """ @@ -19483,14 +19742,13 @@ class HDInsightSparkActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param root_path: Required. The root path in 'sparkJobLinkedService' for all the job’s files. Type: string (or Expression with resultType string). :type root_path: any @@ -19500,12 +19758,10 @@ class HDInsightSparkActivity(ExecutionActivity): :param arguments: The user-specified arguments to HDInsightSparkActivity. :type arguments: list[any] :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.HDInsightActivityDebugInfoOption + :type get_debug_info: str or ~azure.synapse.artifacts.models.HDInsightActivityDebugInfoOption :param spark_job_linked_service: The storage linked service for uploading the entry file and dependencies, and for receiving logs. - :type spark_job_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type spark_job_linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param class_name: The application's Java/Spark main class. :type class_name: str :param proxy_user: The user to impersonate that will execute the job. Type: string (or @@ -19588,22 +19844,19 @@ class HDInsightStreamingActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param storage_linked_services: Storage linked service references. - :type storage_linked_services: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference] + :type storage_linked_services: list[~azure.synapse.artifacts.models.LinkedServiceReference] :param arguments: User specified arguments to HDInsightActivity. :type arguments: list[any] :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.HDInsightActivityDebugInfoOption + :type get_debug_info: str or ~azure.synapse.artifacts.models.HDInsightActivityDebugInfoOption :param mapper: Required. Mapper executable name. Type: string (or Expression with resultType string). :type mapper: any @@ -19617,8 +19870,7 @@ class HDInsightStreamingActivity(ExecutionActivity): :param file_paths: Required. Paths to streaming job files. Can be directories. :type file_paths: list[any] :param file_linked_service: Linked service reference where the files are located. - :type file_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type file_linked_service: ~azure.synapse.artifacts.models.LinkedServiceReference :param combiner: Combiner executable name. Type: string (or Expression with resultType string). :type combiner: any :param command_environment: Command line environment values. @@ -19711,13 +19963,11 @@ class HiveLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. IP address or host name of the Hive server, separated by ';' for @@ -19727,16 +19977,15 @@ class HiveLinkedService(LinkedService): :type port: any :param server_type: The type of Hive server. Possible values include: "HiveServer1", "HiveServer2", "HiveThriftServer". - :type server_type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.HiveServerType + :type server_type: str or ~azure.synapse.artifacts.models.HiveServerType :param thrift_transport_protocol: The transport protocol to use in the Thrift layer. Possible values include: "Binary", "SASL", "HTTP ". :type thrift_transport_protocol: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.HiveThriftTransportProtocol + ~azure.synapse.artifacts.models.HiveThriftTransportProtocol :param authentication_type: Required. The authentication method used to access the Hive server. Possible values include: "Anonymous", "Username", "UsernameAndPassword", "WindowsAzureHDInsightService". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.HiveAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.HiveAuthenticationType :param service_discovery_mode: true to indicate using the ZooKeeper service, false not. :type service_discovery_mode: any :param zoo_keeper_name_space: The namespace on ZooKeeper under which Hive Server 2 nodes are @@ -19749,7 +19998,7 @@ class HiveLinkedService(LinkedService): :type username: any :param password: The password corresponding to the user name that you provided in the Username field. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param http_path: The partial URL corresponding to the Hive server. :type http_path: any :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The @@ -19873,16 +20122,14 @@ class HiveObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -20017,16 +20264,14 @@ class HttpDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param relative_url: The relative URL based on the URL in the HttpLinkedService refers to an HTTP file Type: string (or Expression with resultType string). :type relative_url: any @@ -20043,9 +20288,9 @@ class HttpDataset(Dataset): string). :type additional_headers: any :param format: The format of files. - :type format: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetStorageFormat + :type format: ~azure.synapse.artifacts.models.DatasetStorageFormat :param compression: The data compression method used on files. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression """ _validation = { @@ -20111,13 +20356,11 @@ class HttpLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param url: Required. The base URL of the HTTP endpoint, e.g. http://www.microsoft.com. Type: @@ -20125,14 +20368,13 @@ class HttpLinkedService(LinkedService): :type url: any :param authentication_type: The authentication type to be used to connect to the HTTP server. Possible values include: "Basic", "Anonymous", "Digest", "Windows", "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.HttpAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.HttpAuthenticationType :param user_name: User name for Basic, Digest, or Windows authentication. Type: string (or Expression with resultType string). :type user_name: any :param password: Password for Basic, Digest, Windows, or ClientCertificate with EmbeddedCertData authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param embedded_cert_data: Base64 encoded certificate data for ClientCertificate authentication. For on-premises copy with ClientCertificate authentication, either CertThumbprint or EmbeddedCertData/Password should be specified. Type: string (or Expression @@ -20387,25 +20629,23 @@ class HubspotLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param client_id: Required. The client ID associated with your Hubspot application. :type client_id: any :param client_secret: The client secret associated with your Hubspot application. - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type client_secret: ~azure.synapse.artifacts.models.SecretBase :param access_token: The access token obtained when initially authenticating your OAuth integration. - :type access_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type access_token: ~azure.synapse.artifacts.models.SecretBase :param refresh_token: The refresh token obtained when initially authenticating your OAuth integration. - :type refresh_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type refresh_token: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -20493,16 +20733,14 @@ class HubspotObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -20621,19 +20859,19 @@ class IfConditionActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param expression: Required. An expression that would evaluate to Boolean. This is used to determine the block of activities (ifTrueActivities or ifFalseActivities) that will be executed. - :type expression: ~azure.synapse.artifacts.v2021_06_01_preview.models.Expression + :type expression: ~azure.synapse.artifacts.models.Expression :param if_true_activities: List of activities to execute if expression is evaluated to true. This is an optional property and if not provided, the activity will exit without any action. - :type if_true_activities: list[~azure.synapse.artifacts.v2021_06_01_preview.models.Activity] + :type if_true_activities: list[~azure.synapse.artifacts.models.Activity] :param if_false_activities: List of activities to execute if expression is evaluated to false. This is an optional property and if not provided, the activity will exit without any action. - :type if_false_activities: list[~azure.synapse.artifacts.v2021_06_01_preview.models.Activity] + :type if_false_activities: list[~azure.synapse.artifacts.models.Activity] """ _validation = { @@ -20685,13 +20923,11 @@ class ImpalaLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The IP address or host name of the Impala server. (i.e. @@ -20702,13 +20938,12 @@ class ImpalaLinkedService(LinkedService): :type port: any :param authentication_type: Required. The authentication type to use. Possible values include: "Anonymous", "SASLUsername", "UsernameAndPassword". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.ImpalaAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.ImpalaAuthenticationType :param username: The user name used to access the Impala server. The default value is anonymous when using SASLUsername. :type username: any :param password: The password corresponding to the user name when using UsernameAndPassword. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The default value is false. :type enable_ssl: any @@ -20812,16 +21047,14 @@ class ImpalaObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -20949,13 +21182,11 @@ class InformixLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The non-access credential portion of the connection string @@ -20968,12 +21199,12 @@ class InformixLinkedService(LinkedService): :type authentication_type: any :param credential: The access credential portion of the connection string specified in driver-specific property-value format. - :type credential: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type credential: ~azure.synapse.artifacts.models.SecretBase :param user_name: User name for Basic authentication. Type: string (or Expression with resultType string). :type user_name: any :param password: Password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -21168,16 +21399,14 @@ class InformixTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The Informix table name. Type: string (or Expression with resultType string). :type table_name: any @@ -21233,7 +21462,7 @@ class IntegrationRuntime(msrest.serialization.Model): :type additional_properties: dict[str, any] :param type: Required. Type of integration runtime.Constant filled by server. Possible values include: "Managed", "SelfHosted". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeType + :type type: str or ~azure.synapse.artifacts.models.IntegrationRuntimeType :param description: Integration runtime description. :type description: str """ @@ -21284,10 +21513,9 @@ class IntegrationRuntimeComputeProperties(msrest.serialization.Model): :type max_parallel_executions_per_node: int :param data_flow_properties: Data flow properties for managed integration runtime. :type data_flow_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeDataFlowProperties + ~azure.synapse.artifacts.models.IntegrationRuntimeDataFlowProperties :param v_net_properties: VNet properties for managed integration runtime. - :type v_net_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeVNetProperties + :type v_net_properties: ~azure.synapse.artifacts.models.IntegrationRuntimeVNetProperties """ _validation = { @@ -21334,7 +21562,7 @@ class IntegrationRuntimeCustomSetupScriptProperties(msrest.serialization.Model): script. :type blob_container_uri: str :param sas_token: The SAS token of the Azure blob container. - :type sas_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecureString + :type sas_token: ~azure.synapse.artifacts.models.SecureString """ _attribute_map = { @@ -21362,14 +21590,16 @@ class IntegrationRuntimeDataFlowProperties(msrest.serialization.Model): :type additional_properties: dict[str, any] :param compute_type: Compute type of the cluster which will execute data flow job. Possible values include: "General", "MemoryOptimized", "ComputeOptimized". - :type compute_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowComputeType + :type compute_type: str or ~azure.synapse.artifacts.models.DataFlowComputeType :param core_count: Core count of the cluster which will execute data flow job. Supported values are: 8, 16, 32, 48, 80, 144 and 272. :type core_count: int :param time_to_live: Time to live (in minutes) setting of the cluster which will execute data flow job. :type time_to_live: int + :param cleanup: Cluster will not be recycled and it will be used in next data flow activity run + until TTL (time to live) is reached if this is set as false. Default is true. + :type cleanup: bool """ _validation = { @@ -21381,6 +21611,7 @@ class IntegrationRuntimeDataFlowProperties(msrest.serialization.Model): 'compute_type': {'key': 'computeType', 'type': 'str'}, 'core_count': {'key': 'coreCount', 'type': 'int'}, 'time_to_live': {'key': 'timeToLive', 'type': 'int'}, + 'cleanup': {'key': 'cleanup', 'type': 'bool'}, } def __init__( @@ -21390,6 +21621,7 @@ def __init__( compute_type: Optional[Union[str, "DataFlowComputeType"]] = None, core_count: Optional[int] = None, time_to_live: Optional[int] = None, + cleanup: Optional[bool] = None, **kwargs ): super(IntegrationRuntimeDataFlowProperties, self).__init__(**kwargs) @@ -21397,16 +21629,16 @@ def __init__( self.compute_type = compute_type self.core_count = core_count self.time_to_live = time_to_live + self.cleanup = cleanup class IntegrationRuntimeDataProxyProperties(msrest.serialization.Model): """Data proxy properties for a managed dedicated integration runtime. :param connect_via: The self-hosted integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2021_06_01_preview.models.EntityReference + :type connect_via: ~azure.synapse.artifacts.models.EntityReference :param staging_linked_service: The staging linked service reference. - :type staging_linked_service: - ~azure.synapse.artifacts.v2021_06_01_preview.models.EntityReference + :type staging_linked_service: ~azure.synapse.artifacts.models.EntityReference :param path: The path to contain the staged data in the Blob storage. :type path: str """ @@ -21431,14 +21663,44 @@ def __init__( self.path = path +class IntegrationRuntimeDebugResource(SubResourceDebugResource): + """Integration runtime debug resource. + + All required parameters must be populated in order to send to Azure. + + :param name: The resource name. + :type name: str + :param properties: Required. Integration runtime properties. + :type properties: ~azure.synapse.artifacts.models.IntegrationRuntime + """ + + _validation = { + 'properties': {'required': True}, + } + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'properties': {'key': 'properties', 'type': 'IntegrationRuntime'}, + } + + def __init__( + self, + *, + properties: "IntegrationRuntime", + name: Optional[str] = None, + **kwargs + ): + super(IntegrationRuntimeDebugResource, self).__init__(name=name, **kwargs) + self.properties = properties + + class IntegrationRuntimeListResponse(msrest.serialization.Model): """A list of integration runtime resources. All required parameters must be populated in order to send to Azure. :param value: Required. List of integration runtimes. - :type value: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeResource] + :type value: list[~azure.synapse.artifacts.models.IntegrationRuntimeResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -21471,8 +21733,7 @@ class IntegrationRuntimeReference(msrest.serialization.Model): :param type: Required. Type of integration runtime. Possible values include: "IntegrationRuntimeReference". - :type type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReferenceType + :type type: str or ~azure.synapse.artifacts.models.IntegrationRuntimeReferenceType :param reference_name: Required. Reference integration runtime name. :type reference_name: str :param parameters: Arguments for integration runtime. @@ -21522,7 +21783,7 @@ class IntegrationRuntimeResource(SubResource): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Integration runtime properties. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntime + :type properties: ~azure.synapse.artifacts.models.IntegrationRuntime """ _validation = { @@ -21563,12 +21824,12 @@ class IntegrationRuntimeSsisCatalogInfo(msrest.serialization.Model): :type catalog_admin_user_name: str :param catalog_admin_password: The password of the administrator user account of the catalog database. - :type catalog_admin_password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecureString + :type catalog_admin_password: ~azure.synapse.artifacts.models.SecureString :param catalog_pricing_tier: The pricing tier for the catalog database. The valid values could be found in https://azure.microsoft.com/en-us/pricing/details/sql-database/. Possible values include: "Basic", "Standard", "Premium", "PremiumRS". :type catalog_pricing_tier: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeSsisCatalogPricingTier + ~azure.synapse.artifacts.models.IntegrationRuntimeSsisCatalogPricingTier """ _validation = { @@ -21608,28 +21869,24 @@ class IntegrationRuntimeSsisProperties(msrest.serialization.Model): collection. :type additional_properties: dict[str, any] :param catalog_info: Catalog information for managed dedicated integration runtime. - :type catalog_info: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeSsisCatalogInfo + :type catalog_info: ~azure.synapse.artifacts.models.IntegrationRuntimeSsisCatalogInfo :param license_type: License type for bringing your own license scenario. Possible values include: "BasePrice", "LicenseIncluded". - :type license_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeLicenseType + :type license_type: str or ~azure.synapse.artifacts.models.IntegrationRuntimeLicenseType :param custom_setup_script_properties: Custom setup script properties for a managed dedicated integration runtime. :type custom_setup_script_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeCustomSetupScriptProperties + ~azure.synapse.artifacts.models.IntegrationRuntimeCustomSetupScriptProperties :param data_proxy_properties: Data proxy properties for a managed dedicated integration runtime. :type data_proxy_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeDataProxyProperties + ~azure.synapse.artifacts.models.IntegrationRuntimeDataProxyProperties :param edition: The edition for the SSIS Integration Runtime. Possible values include: "Standard", "Enterprise". - :type edition: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeEdition + :type edition: str or ~azure.synapse.artifacts.models.IntegrationRuntimeEdition :param express_custom_setup_properties: Custom setup without script properties for a SSIS integration runtime. - :type express_custom_setup_properties: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.CustomSetupBase] + :type express_custom_setup_properties: list[~azure.synapse.artifacts.models.CustomSetupBase] """ _attribute_map = { @@ -21713,13 +21970,11 @@ class JiraLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The IP address or host name of the Jira service. (e.g. @@ -21732,7 +21987,7 @@ class JiraLinkedService(LinkedService): :type username: any :param password: The password corresponding to the user name that you provided in the username field. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -21821,16 +22076,14 @@ class JiraObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -21953,18 +22206,16 @@ class JsonDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param location: The location of the json data storage. - :type location: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetLocation + :type location: ~azure.synapse.artifacts.models.DatasetLocation :param encoding_name: The code page name of the preferred encoding. If not specified, the default value is UTF-8, unless BOM denotes another Unicode encoding. Refer to the name column of the table in the following link to set supported values: @@ -21972,7 +22223,7 @@ class JsonDataset(Dataset): resultType string). :type encoding_name: any :param compression: The data compression method used for the json dataset. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression """ _validation = { @@ -22033,10 +22284,8 @@ class JsonFormat(DatasetStorageFormat): :param deserializer: Deserializer. Type: string (or Expression with resultType string). :type deserializer: any :param file_pattern: File pattern of JSON. To be more specific, the way of separating a - collection of JSON objects. The default value is 'setOfObjects'. It is case-sensitive. Possible - values include: "setOfObjects", "arrayOfObjects". - :type file_pattern: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.JsonFormatFilePattern + collection of JSON objects. The default value is 'setOfObjects'. It is case-sensitive. + :type file_pattern: any :param nesting_separator: The character used to separate nesting levels. Default value is '.' (dot). Type: string (or Expression with resultType string). :type nesting_separator: any @@ -22066,7 +22315,7 @@ class JsonFormat(DatasetStorageFormat): 'type': {'key': 'type', 'type': 'str'}, 'serializer': {'key': 'serializer', 'type': 'object'}, 'deserializer': {'key': 'deserializer', 'type': 'object'}, - 'file_pattern': {'key': 'filePattern', 'type': 'str'}, + 'file_pattern': {'key': 'filePattern', 'type': 'object'}, 'nesting_separator': {'key': 'nestingSeparator', 'type': 'object'}, 'encoding_name': {'key': 'encodingName', 'type': 'object'}, 'json_node_reference': {'key': 'jsonNodeReference', 'type': 'object'}, @@ -22079,7 +22328,7 @@ def __init__( additional_properties: Optional[Dict[str, Any]] = None, serializer: Optional[Any] = None, deserializer: Optional[Any] = None, - file_pattern: Optional[Union[str, "JsonFormatFilePattern"]] = None, + file_pattern: Optional[Any] = None, nesting_separator: Optional[Any] = None, encoding_name: Optional[Any] = None, json_node_reference: Optional[Any] = None, @@ -22106,8 +22355,7 @@ class JsonReadSettings(FormatReadSettings): :param type: Required. The read setting type.Constant filled by server. :type type: str :param compression_properties: Compression settings. - :type compression_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.CompressionReadSettings + :type compression_properties: ~azure.synapse.artifacts.models.CompressionReadSettings """ _validation = { @@ -22158,9 +22406,9 @@ class JsonSink(CopySink): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Json store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreWriteSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreWriteSettings :param format_settings: Json format settings. - :type format_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.JsonWriteSettings + :type format_settings: ~azure.synapse.artifacts.models.JsonWriteSettings """ _validation = { @@ -22218,9 +22466,9 @@ class JsonSource(CopySource): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Json store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings :param format_settings: Json format settings. - :type format_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.JsonReadSettings + :type format_settings: ~azure.synapse.artifacts.models.JsonReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :type additional_columns: any @@ -22271,10 +22519,8 @@ class JsonWriteSettings(FormatWriteSettings): :param type: Required. The write setting type.Constant filled by server. :type type: str :param file_pattern: File pattern of JSON. This setting controls the way a collection of JSON - objects will be treated. The default value is 'setOfObjects'. It is case-sensitive. Possible - values include: "setOfObjects", "arrayOfObjects". - :type file_pattern: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.JsonWriteFilePattern + objects will be treated. The default value is 'setOfObjects'. It is case-sensitive. + :type file_pattern: any """ _validation = { @@ -22284,14 +22530,14 @@ class JsonWriteSettings(FormatWriteSettings): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, 'type': {'key': 'type', 'type': 'str'}, - 'file_pattern': {'key': 'filePattern', 'type': 'str'}, + 'file_pattern': {'key': 'filePattern', 'type': 'object'}, } def __init__( self, *, additional_properties: Optional[Dict[str, Any]] = None, - file_pattern: Optional[Union[str, "JsonWriteFilePattern"]] = None, + file_pattern: Optional[Any] = None, **kwargs ): super(JsonWriteSettings, self).__init__(additional_properties=additional_properties, **kwargs) @@ -22299,6 +22545,167 @@ def __init__( self.file_pattern = file_pattern +class KqlScript(msrest.serialization.Model): + """KqlScript. + + :param content: + :type content: ~azure.synapse.artifacts.models.KqlScriptContent + """ + + _attribute_map = { + 'content': {'key': 'content', 'type': 'KqlScriptContent'}, + } + + def __init__( + self, + *, + content: Optional["KqlScriptContent"] = None, + **kwargs + ): + super(KqlScript, self).__init__(**kwargs) + self.content = content + + +class KqlScriptContent(msrest.serialization.Model): + """KqlScriptContent. + + :param query: + :type query: str + :param metadata: + :type metadata: ~azure.synapse.artifacts.models.KqlScriptContentMetadata + :param current_connection: + :type current_connection: ~azure.synapse.artifacts.models.KqlScriptContentCurrentConnection + """ + + _attribute_map = { + 'query': {'key': 'query', 'type': 'str'}, + 'metadata': {'key': 'metadata', 'type': 'KqlScriptContentMetadata'}, + 'current_connection': {'key': 'currentConnection', 'type': 'KqlScriptContentCurrentConnection'}, + } + + def __init__( + self, + *, + query: Optional[str] = None, + metadata: Optional["KqlScriptContentMetadata"] = None, + current_connection: Optional["KqlScriptContentCurrentConnection"] = None, + **kwargs + ): + super(KqlScriptContent, self).__init__(**kwargs) + self.query = query + self.metadata = metadata + self.current_connection = current_connection + + +class KqlScriptContentCurrentConnection(msrest.serialization.Model): + """KqlScriptContentCurrentConnection. + + :param name: + :type name: str + :param type: + :type type: str + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + def __init__( + self, + *, + name: Optional[str] = None, + type: Optional[str] = None, + **kwargs + ): + super(KqlScriptContentCurrentConnection, self).__init__(**kwargs) + self.name = name + self.type = type + + +class KqlScriptContentMetadata(msrest.serialization.Model): + """KqlScriptContentMetadata. + + :param language: + :type language: str + """ + + _attribute_map = { + 'language': {'key': 'language', 'type': 'str'}, + } + + def __init__( + self, + *, + language: Optional[str] = None, + **kwargs + ): + super(KqlScriptContentMetadata, self).__init__(**kwargs) + self.language = language + + +class KqlScriptResource(msrest.serialization.Model): + """KqlScriptResource. + + :param id: + :type id: str + :param name: + :type name: str + :param type: + :type type: str + :param properties: Properties of sql script. + :type properties: ~azure.synapse.artifacts.models.KqlScript + """ + + _attribute_map = { + 'id': {'key': 'id', 'type': 'str'}, + 'name': {'key': 'name', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'properties': {'key': 'properties', 'type': 'KqlScript'}, + } + + def __init__( + self, + *, + id: Optional[str] = None, + name: Optional[str] = None, + type: Optional[str] = None, + properties: Optional["KqlScript"] = None, + **kwargs + ): + super(KqlScriptResource, self).__init__(**kwargs) + self.id = id + self.name = name + self.type = type + self.properties = properties + + +class KqlScriptsResourceCollectionResponse(msrest.serialization.Model): + """KqlScriptsResourceCollectionResponse. + + :param value: + :type value: list[~azure.synapse.artifacts.models.KqlScriptResource] + :param next_link: + :type next_link: str + """ + + _attribute_map = { + 'value': {'key': 'value', 'type': '[KqlScriptResource]'}, + 'next_link': {'key': 'nextLink', 'type': 'str'}, + } + + def __init__( + self, + *, + value: Optional[List["KqlScriptResource"]] = None, + next_link: Optional[str] = None, + **kwargs + ): + super(KqlScriptsResourceCollectionResponse, self).__init__(**kwargs) + self.value = value + self.next_link = next_link + + class LibraryInfo(msrest.serialization.Model): """Library/package information of a Big Data pool powered by Apache Spark. @@ -22310,8 +22717,8 @@ class LibraryInfo(msrest.serialization.Model): :type path: str :param container_name: Storage blob container name. :type container_name: str - :ivar uploaded_timestamp: The last update time of the library. - :vartype uploaded_timestamp: ~datetime.datetime + :param uploaded_timestamp: The last update time of the library. + :type uploaded_timestamp: ~datetime.datetime :param type: Type of the library. :type type: str :ivar provisioning_status: Provisioning status of the library/package. @@ -22321,7 +22728,6 @@ class LibraryInfo(msrest.serialization.Model): """ _validation = { - 'uploaded_timestamp': {'readonly': True}, 'provisioning_status': {'readonly': True}, 'creator_id': {'readonly': True}, } @@ -22342,6 +22748,7 @@ def __init__( name: Optional[str] = None, path: Optional[str] = None, container_name: Optional[str] = None, + uploaded_timestamp: Optional[datetime.datetime] = None, type: Optional[str] = None, **kwargs ): @@ -22349,7 +22756,7 @@ def __init__( self.name = name self.path = path self.container_name = container_name - self.uploaded_timestamp = None + self.uploaded_timestamp = uploaded_timestamp self.type = type self.provisioning_status = None self.creator_id = None @@ -22361,7 +22768,7 @@ class LibraryListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of Library. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.LibraryResource] + :type value: list[~azure.synapse.artifacts.models.LibraryResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -22441,7 +22848,7 @@ class LibraryResource(SubResource): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Library/package properties. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.LibraryResourceProperties + :type properties: ~azure.synapse.artifacts.models.LibraryResourceProperties """ _validation = { @@ -22633,7 +23040,7 @@ class LinkedIntegrationRuntimeKeyAuthorization(LinkedIntegrationRuntimeType): sharing.Constant filled by server. :type authorization_type: str :param key: Required. The key used for authorization. - :type key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecureString + :type key: ~azure.synapse.artifacts.models.SecureString """ _validation = { @@ -22698,7 +23105,7 @@ class LinkedServiceDebugResource(SubResourceDebugResource): :param name: The resource name. :type name: str :param properties: Required. Properties of linked service. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedService + :type properties: ~azure.synapse.artifacts.models.LinkedService """ _validation = { @@ -22727,7 +23134,7 @@ class LinkedServiceListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of linked services. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceResource] + :type value: list[~azure.synapse.artifacts.models.LinkedServiceResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -22760,7 +23167,7 @@ class LinkedServiceReference(msrest.serialization.Model): :param type: Required. Linked service reference type. Possible values include: "LinkedServiceReference". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.Type + :type type: str or ~azure.synapse.artifacts.models.Type :param reference_name: Required. Reference LinkedService name. :type reference_name: str :param parameters: Arguments for LinkedService. @@ -22810,7 +23217,7 @@ class LinkedServiceResource(SubResource): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Properties of linked service. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedService + :type properties: ~azure.synapse.artifacts.models.LinkedService """ _validation = { @@ -22845,8 +23252,7 @@ class LogLocationSettings(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param linked_service_name: Required. Log storage linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param path: The path to storage for storing detailed logs of activity execution. Type: string (or Expression with resultType string). :type path: any @@ -22882,12 +23288,10 @@ class LogSettings(msrest.serialization.Model): (or Expression with resultType boolean). :type enable_copy_activity_log: any :param copy_activity_log_settings: Specifies settings for copy activity log. - :type copy_activity_log_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.CopyActivityLogSettings + :type copy_activity_log_settings: ~azure.synapse.artifacts.models.CopyActivityLogSettings :param log_location_settings: Required. Log location settings customer needs to provide when enabling log. - :type log_location_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LogLocationSettings + :type log_location_settings: ~azure.synapse.artifacts.models.LogLocationSettings """ _validation = { @@ -22923,8 +23327,7 @@ class LogStorageSettings(msrest.serialization.Model): collection. :type additional_properties: dict[str, any] :param linked_service_name: Required. Log storage linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param path: The path to storage for storing detailed logs of activity execution. Type: string (or Expression with resultType string). :type path: any @@ -22981,18 +23384,17 @@ class LookupActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param source: Required. Dataset-specific source properties, same as copy activity source. - :type source: ~azure.synapse.artifacts.v2021_06_01_preview.models.CopySource + :type source: ~azure.synapse.artifacts.models.CopySource :param dataset: Required. Lookup activity dataset reference. - :type dataset: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference + :type dataset: ~azure.synapse.artifacts.models.DatasetReference :param first_row_only: Whether to return first row or all rows. Default value is true. Type: boolean (or Expression with resultType boolean). :type first_row_only: any @@ -23052,19 +23454,17 @@ class MagentoLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The URL of the Magento instance. (i.e. 192.168.222.110/magento3). :type host: any :param access_token: The access token from Magento. - :type access_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type access_token: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -23146,16 +23546,14 @@ class MagentoObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -23270,7 +23668,7 @@ class ManagedIdentity(msrest.serialization.Model): :vartype tenant_id: str :param type: The type of managed identity for the workspace. Possible values include: "None", "SystemAssigned". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.ResourceIdentityType + :type type: str or ~azure.synapse.artifacts.models.ResourceIdentityType """ _validation = { @@ -23308,23 +23706,19 @@ class ManagedIntegrationRuntime(IntegrationRuntime): :type additional_properties: dict[str, any] :param type: Required. Type of integration runtime.Constant filled by server. Possible values include: "Managed", "SelfHosted". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeType + :type type: str or ~azure.synapse.artifacts.models.IntegrationRuntimeType :param description: Integration runtime description. :type description: str :ivar state: Integration runtime state, only valid for managed dedicated integration runtime. Possible values include: "Initial", "Stopped", "Started", "Starting", "Stopping", "NeedRegistration", "Online", "Limited", "Offline", "AccessDenied". - :vartype state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeState + :vartype state: str or ~azure.synapse.artifacts.models.IntegrationRuntimeState :param managed_virtual_network: Managed Virtual Network reference. - :type managed_virtual_network: - ~azure.synapse.artifacts.v2021_06_01_preview.models.ManagedVirtualNetworkReference + :type managed_virtual_network: ~azure.synapse.artifacts.models.ManagedVirtualNetworkReference :param compute_properties: The compute resource for managed integration runtime. - :type compute_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeComputeProperties + :type compute_properties: ~azure.synapse.artifacts.models.IntegrationRuntimeComputeProperties :param ssis_properties: SSIS properties for managed integration runtime. - :type ssis_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeSsisProperties + :type ssis_properties: ~azure.synapse.artifacts.models.IntegrationRuntimeSsisProperties """ _validation = { @@ -23440,13 +23834,13 @@ class MappingDataFlow(DataFlow): :type annotations: list[any] :param folder: The folder that this data flow is in. If not specified, Data flow will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowFolder + :type folder: ~azure.synapse.artifacts.models.DataFlowFolder :param sources: List of sources in data flow. - :type sources: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowSource] + :type sources: list[~azure.synapse.artifacts.models.DataFlowSource] :param sinks: List of sinks in data flow. - :type sinks: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowSink] + :type sinks: list[~azure.synapse.artifacts.models.DataFlowSink] :param transformations: List of transformations in data flow. - :type transformations: list[~azure.synapse.artifacts.v2021_06_01_preview.models.Transformation] + :type transformations: list[~azure.synapse.artifacts.models.Transformation] :param script: DataFlow script. :type script: str """ @@ -23497,20 +23891,18 @@ class MariaDBLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type pwd: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -23634,16 +24026,14 @@ class MariaDBTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -23696,13 +24086,11 @@ class MarketoLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param endpoint: Required. The endpoint of the Marketo server. (i.e. 123-ABC-321.mktorest.com). @@ -23710,7 +24098,7 @@ class MarketoLinkedService(LinkedService): :param client_id: Required. The client Id of your Marketo service. :type client_id: any :param client_secret: The client secret of your Marketo service. - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type client_secret: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -23796,16 +24184,14 @@ class MarketoObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -23920,13 +24306,11 @@ class MicrosoftAccessLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The non-access credential portion of the connection string @@ -23939,12 +24323,12 @@ class MicrosoftAccessLinkedService(LinkedService): :type authentication_type: any :param credential: The access credential portion of the connection string specified in driver-specific property-value format. - :type credential: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type credential: ~azure.synapse.artifacts.models.SecretBase :param user_name: User name for Basic authentication. Type: string (or Expression with resultType string). :type user_name: any :param password: Password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -24135,16 +24519,14 @@ class MicrosoftAccessTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The Microsoft Access table name. Type: string (or Expression with resultType string). :type table_name: any @@ -24206,16 +24588,14 @@ class MongoDbAtlasCollectionDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param collection: Required. The collection name of the MongoDB Atlas database. Type: string (or Expression with resultType string). :type collection: any @@ -24270,13 +24650,11 @@ class MongoDbAtlasLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The MongoDB Atlas connection string. Type: string, @@ -24347,8 +24725,7 @@ class MongoDbAtlasSource(CopySource): with resultType string). :type filter: any :param cursor_methods: Cursor methods for Mongodb query. - :type cursor_methods: - ~azure.synapse.artifacts.v2021_06_01_preview.models.MongoDbCursorMethodsProperties + :type cursor_methods: ~azure.synapse.artifacts.models.MongoDbCursorMethodsProperties :param batch_size: Specifies the number of documents to return in each batch of the response from MongoDB Atlas instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response @@ -24421,16 +24798,14 @@ class MongoDbCollectionDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param collection_name: Required. The table name of the MongoDB database. Type: string (or Expression with resultType string). :type collection_name: any @@ -24534,13 +24909,11 @@ class MongoDbLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param server: Required. The IP address or server name of the MongoDB server. Type: string (or @@ -24548,8 +24921,7 @@ class MongoDbLinkedService(LinkedService): :type server: any :param authentication_type: The authentication type to be used to connect to the MongoDB database. Possible values include: "Basic", "Anonymous". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.MongoDbAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.MongoDbAuthenticationType :param database_name: Required. The name of the MongoDB database that you want to access. Type: string (or Expression with resultType string). :type database_name: any @@ -24557,7 +24929,7 @@ class MongoDbLinkedService(LinkedService): string). :type username: any :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param auth_source: Database to verify the username and password. Type: string (or Expression with resultType string). :type auth_source: any @@ -24712,16 +25084,14 @@ class MongoDbV2CollectionDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param collection: Required. The collection name of the MongoDB database. Type: string (or Expression with resultType string). :type collection: any @@ -24776,13 +25146,11 @@ class MongoDbV2LinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The MongoDB connection string. Type: string, SecureString @@ -24852,8 +25220,7 @@ class MongoDbV2Source(CopySource): with resultType string). :type filter: any :param cursor_methods: Cursor methods for Mongodb query. - :type cursor_methods: - ~azure.synapse.artifacts.v2021_06_01_preview.models.MongoDbCursorMethodsProperties + :type cursor_methods: ~azure.synapse.artifacts.models.MongoDbCursorMethodsProperties :param batch_size: Specifies the number of documents to return in each batch of the response from MongoDB instance. In most cases, modifying the batch size will not affect the user or the application. This property's main purpose is to avoid hit the limitation of response size. @@ -24918,20 +25285,17 @@ class MySqlLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The connection string. :type connection_string: any :param password: The Azure key vault secret reference of password in connection string. - :type password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type password: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -25055,16 +25419,14 @@ class MySqlTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The MySQL table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -25117,20 +25479,18 @@ class NetezzaLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type pwd: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -25239,11 +25599,9 @@ class NetezzaSource(TabularSource): :type query: any :param partition_option: The partition mechanism that will be used for Netezza read in parallel. Possible values include: "None", "DataSlice", "DynamicRange". - :type partition_option: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.NetezzaPartitionOption + :type partition_option: str or ~azure.synapse.artifacts.models.NetezzaPartitionOption :param partition_settings: The settings that will be leveraged for Netezza source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.NetezzaPartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.NetezzaPartitionSettings """ _validation = { @@ -25303,16 +25661,14 @@ class NetezzaTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -25378,12 +25734,11 @@ class Notebook(msrest.serialization.Model): :param description: The description of the notebook. :type description: str :param big_data_pool: Big data pool reference. - :type big_data_pool: ~azure.synapse.artifacts.v2021_06_01_preview.models.BigDataPoolReference + :type big_data_pool: ~azure.synapse.artifacts.models.BigDataPoolReference :param session_properties: Session properties. - :type session_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookSessionProperties + :type session_properties: ~azure.synapse.artifacts.models.NotebookSessionProperties :param metadata: Required. Notebook root-level metadata. - :type metadata: ~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookMetadata + :type metadata: ~azure.synapse.artifacts.models.NotebookMetadata :param nbformat: Required. Notebook format (major number). Incremented between backwards incompatible changes to the notebook format. :type nbformat: int @@ -25391,7 +25746,10 @@ class Notebook(msrest.serialization.Model): compatible changes to the notebook format. :type nbformat_minor: int :param cells: Required. Array of cells of the current notebook. - :type cells: list[~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookCell] + :type cells: list[~azure.synapse.artifacts.models.NotebookCell] + :param folder: The folder that this notebook is in. If not specified, this notebook will appear + at the root level. + :type folder: ~azure.synapse.artifacts.models.NotebookFolder """ _validation = { @@ -25410,6 +25768,7 @@ class Notebook(msrest.serialization.Model): 'nbformat': {'key': 'nbformat', 'type': 'int'}, 'nbformat_minor': {'key': 'nbformat_minor', 'type': 'int'}, 'cells': {'key': 'cells', 'type': '[NotebookCell]'}, + 'folder': {'key': 'folder', 'type': 'NotebookFolder'}, } def __init__( @@ -25423,6 +25782,7 @@ def __init__( description: Optional[str] = None, big_data_pool: Optional["BigDataPoolReference"] = None, session_properties: Optional["NotebookSessionProperties"] = None, + folder: Optional["NotebookFolder"] = None, **kwargs ): super(Notebook, self).__init__(**kwargs) @@ -25434,6 +25794,7 @@ def __init__( self.nbformat = nbformat self.nbformat_minor = nbformat_minor self.cells = cells + self.folder = folder class NotebookCell(msrest.serialization.Model): @@ -25453,7 +25814,7 @@ class NotebookCell(msrest.serialization.Model): :param attachments: Attachments associated with the cell. :type attachments: any :param outputs: Cell-level output items. - :type outputs: list[~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookCellOutputItem] + :type outputs: list[~azure.synapse.artifacts.models.NotebookCellOutputItem] """ _validation = { @@ -25502,7 +25863,7 @@ class NotebookCellOutputItem(msrest.serialization.Model): :type execution_count: int :param output_type: Required. Execution, display, or stream outputs. Possible values include: "execute_result", "display_data", "stream", "error". - :type output_type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.CellOutputType + :type output_type: str or ~azure.synapse.artifacts.models.CellOutputType :param text: For output_type=stream, the stream's text output, represented as a string or an array of strings. :type text: any @@ -25545,6 +25906,27 @@ def __init__( self.metadata = metadata +class NotebookFolder(msrest.serialization.Model): + """The folder that this notebook is in. If not specified, this notebook will appear at the root level. + + :param name: The name of the folder that this notebook is in. + :type name: str + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + } + + def __init__( + self, + *, + name: Optional[str] = None, + **kwargs + ): + super(NotebookFolder, self).__init__(**kwargs) + self.name = name + + class NotebookKernelSpec(msrest.serialization.Model): """Kernel information. @@ -25628,7 +26010,7 @@ class NotebookListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of Notebooks. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookResource] + :type value: list[~azure.synapse.artifacts.models.NotebookResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -25661,9 +26043,9 @@ class NotebookMetadata(msrest.serialization.Model): collection. :type additional_properties: dict[str, any] :param kernelspec: Kernel information. - :type kernelspec: ~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookKernelSpec + :type kernelspec: ~azure.synapse.artifacts.models.NotebookKernelSpec :param language_info: Language info. - :type language_info: ~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookLanguageInfo + :type language_info: ~azure.synapse.artifacts.models.NotebookLanguageInfo """ _attribute_map = { @@ -25704,7 +26086,7 @@ class NotebookResource(msrest.serialization.Model): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Properties of Notebook. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.Notebook + :type properties: ~azure.synapse.artifacts.models.Notebook """ _validation = { @@ -25800,13 +26182,11 @@ class ODataLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param url: Required. The URL of the OData service endpoint. Type: string (or Expression with @@ -25815,13 +26195,12 @@ class ODataLinkedService(LinkedService): :param authentication_type: Type of authentication used to connect to the OData service. Possible values include: "Basic", "Anonymous", "Windows", "AadServicePrincipal", "ManagedServiceIdentity". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.ODataAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.ODataAuthenticationType :param user_name: User name of the OData service. Type: string (or Expression with resultType string). :type user_name: any :param password: Password of the OData service. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param tenant: Specify the tenant information (domain name or tenant ID) under which your application resides. Type: string (or Expression with resultType string). :type tenant: any @@ -25838,20 +26217,18 @@ class ODataLinkedService(LinkedService): :param aad_service_principal_credential_type: Specify the credential type (key or cert) is used for service principal. Possible values include: "ServicePrincipalKey", "ServicePrincipalCert". :type aad_service_principal_credential_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.ODataAadServicePrincipalCredentialType + ~azure.synapse.artifacts.models.ODataAadServicePrincipalCredentialType :param service_principal_key: Specify the secret of your application registered in Azure Active Directory. Type: string (or Expression with resultType string). - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param service_principal_embedded_cert: Specify the base64 encoded certificate of your application registered in Azure Active Directory. Type: string (or Expression with resultType string). - :type service_principal_embedded_cert: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_embedded_cert: ~azure.synapse.artifacts.models.SecretBase :param service_principal_embedded_cert_password: Specify the password of your certificate if your certificate has a password and you are using AadServicePrincipal authentication. Type: string (or Expression with resultType string). - :type service_principal_embedded_cert_password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_embedded_cert_password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -25944,16 +26321,14 @@ class ODataResourceDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param path: The OData resource path. Type: string (or Expression with resultType string). :type path: any """ @@ -26072,13 +26447,11 @@ class OdbcLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The non-access credential portion of the connection string @@ -26090,12 +26463,12 @@ class OdbcLinkedService(LinkedService): :type authentication_type: any :param credential: The access credential portion of the connection string specified in driver-specific property-value format. - :type credential: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type credential: ~azure.synapse.artifacts.models.SecretBase :param user_name: User name for Basic authentication. Type: string (or Expression with resultType string). :type user_name: any :param password: Password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -26290,16 +26663,14 @@ class OdbcTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The ODBC table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -26360,16 +26731,14 @@ class Office365Dataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: Required. Name of the dataset to extract from Office 365. Type: string (or Expression with resultType string). :type table_name: any @@ -26430,13 +26799,11 @@ class Office365LinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param office365_tenant_id: Required. Azure tenant ID to which the Office 365 account belongs. @@ -26449,7 +26816,7 @@ class Office365LinkedService(LinkedService): Expression with resultType string). :type service_principal_id: any :param service_principal_key: Required. Specify the application's key. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -26599,7 +26966,7 @@ class OperationResult(msrest.serialization.Model): :param target: Property name/path in request associated with error. :type target: str :param details: Array with additional error details. - :type details: list[~azure.synapse.artifacts.v2021_06_01_preview.models.CloudError] + :type details: list[~azure.synapse.artifacts.models.CloudErrorAutoGenerated] """ _validation = { @@ -26611,7 +26978,7 @@ class OperationResult(msrest.serialization.Model): 'code': {'key': 'error.code', 'type': 'str'}, 'message': {'key': 'error.message', 'type': 'str'}, 'target': {'key': 'error.target', 'type': 'str'}, - 'details': {'key': 'error.details', 'type': '[CloudError]'}, + 'details': {'key': 'error.details', 'type': '[CloudErrorAutoGenerated]'}, } def __init__( @@ -26620,7 +26987,7 @@ def __init__( code: Optional[str] = None, message: Optional[str] = None, target: Optional[str] = None, - details: Optional[List["CloudError"]] = None, + details: Optional[List["CloudErrorAutoGenerated"]] = None, **kwargs ): super(OperationResult, self).__init__(**kwargs) @@ -26642,21 +27009,18 @@ class OracleLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param password: The Azure key vault secret reference of password in connection string. - :type password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type password: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -26752,13 +27116,11 @@ class OracleServiceCloudLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The URL of the Oracle Service Cloud instance. @@ -26767,7 +27129,7 @@ class OracleServiceCloudLinkedService(LinkedService): :type username: any :param password: Required. The password corresponding to the user name that you provided in the username key. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. Type: boolean (or Expression with resultType boolean). :type use_encrypted_endpoints: any @@ -26855,16 +27217,14 @@ class OracleServiceCloudObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -27057,11 +27417,9 @@ class OracleSource(CopySource): :type query_timeout: any :param partition_option: The partition mechanism that will be used for Oracle read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". - :type partition_option: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.OraclePartitionOption + :type partition_option: str or ~azure.synapse.artifacts.models.OraclePartitionOption :param partition_settings: The settings that will be leveraged for Oracle source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.OraclePartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.OraclePartitionSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :type additional_columns: any @@ -27126,16 +27484,14 @@ class OracleTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -27209,21 +27565,19 @@ class OrcDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param location: The location of the ORC data storage. - :type location: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetLocation - :param orc_compression_codec: Possible values include: "none", "zlib", "snappy", "lzo". - :type orc_compression_codec: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.OrcCompressionCodec + :type location: ~azure.synapse.artifacts.models.DatasetLocation + :param orc_compression_codec: The data orcCompressionCodec. Type: string (or Expression with + resultType string). + :type orc_compression_codec: any """ _validation = { @@ -27242,7 +27596,7 @@ class OrcDataset(Dataset): 'annotations': {'key': 'annotations', 'type': '[object]'}, 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, 'location': {'key': 'typeProperties.location', 'type': 'DatasetLocation'}, - 'orc_compression_codec': {'key': 'typeProperties.orcCompressionCodec', 'type': 'str'}, + 'orc_compression_codec': {'key': 'typeProperties.orcCompressionCodec', 'type': 'object'}, } def __init__( @@ -27257,7 +27611,7 @@ def __init__( annotations: Optional[List[Any]] = None, folder: Optional["DatasetFolder"] = None, location: Optional["DatasetLocation"] = None, - orc_compression_codec: Optional[Union[str, "OrcCompressionCodec"]] = None, + orc_compression_codec: Optional[Any] = None, **kwargs ): super(OrcDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) @@ -27331,9 +27685,9 @@ class OrcSink(CopySink): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: ORC store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreWriteSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreWriteSettings :param format_settings: ORC format settings. - :type format_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.OrcWriteSettings + :type format_settings: ~azure.synapse.artifacts.models.OrcWriteSettings """ _validation = { @@ -27391,7 +27745,7 @@ class OrcSource(CopySource): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: ORC store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :type additional_columns: any @@ -27479,7 +27833,7 @@ class ParameterSpecification(msrest.serialization.Model): :param type: Required. Parameter type. Possible values include: "Object", "String", "Int", "Float", "Bool", "Array", "SecureString". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterType + :type type: str or ~azure.synapse.artifacts.models.ParameterType :param default_value: Default value of parameter. :type default_value: any """ @@ -27524,18 +27878,16 @@ class ParquetDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param location: The location of the parquet storage. - :type location: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetLocation + :type location: ~azure.synapse.artifacts.models.DatasetLocation :param compression_codec: A string from ParquetCompressionCodecEnum or an expression. :type compression_codec: any """ @@ -27645,9 +27997,9 @@ class ParquetSink(CopySink): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Parquet store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreWriteSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreWriteSettings :param format_settings: Parquet format settings. - :type format_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.ParquetWriteSettings + :type format_settings: ~azure.synapse.artifacts.models.ParquetWriteSettings """ _validation = { @@ -27705,7 +28057,7 @@ class ParquetSource(CopySource): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Parquet store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :type additional_columns: any @@ -27797,13 +28149,11 @@ class PaypalLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The URL of the PayPal instance. (i.e. api.sandbox.paypal.com). @@ -27811,7 +28161,7 @@ class PaypalLinkedService(LinkedService): :param client_id: Required. The client ID associated with your PayPal application. :type client_id: any :param client_secret: The client secret associated with your PayPal application. - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type client_secret: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -27897,16 +28247,14 @@ class PaypalObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -28021,13 +28369,11 @@ class PhoenixLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The IP address or host name of the Phoenix server. (i.e. @@ -28043,12 +28389,11 @@ class PhoenixLinkedService(LinkedService): :param authentication_type: Required. The authentication mechanism used to connect to the Phoenix server. Possible values include: "Anonymous", "UsernameAndPassword", "WindowsAzureHDInsightService". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.PhoenixAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.PhoenixAuthenticationType :param username: The user name used to connect to the Phoenix server. :type username: any :param password: The password corresponding to the user name. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The default value is false. :type enable_ssl: any @@ -28155,16 +28500,14 @@ class PhoenixObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -28308,7 +28651,7 @@ class PipelineListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of pipelines. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineResource] + :type value: list[~azure.synapse.artifacts.models.PipelineResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -28340,7 +28683,7 @@ class PipelineReference(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param type: Required. Pipeline reference type. Possible values include: "PipelineReference". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineReferenceType + :type type: str or ~azure.synapse.artifacts.models.PipelineReferenceType :param reference_name: Required. Reference pipeline name. :type reference_name: str :param name: Reference name. @@ -28393,13 +28736,11 @@ class PipelineResource(SubResource): :param description: The description of the pipeline. :type description: str :param activities: List of activities in pipeline. - :type activities: list[~azure.synapse.artifacts.v2021_06_01_preview.models.Activity] + :type activities: list[~azure.synapse.artifacts.models.Activity] :param parameters: List of parameters for pipeline. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param variables: List of variables for pipeline. - :type variables: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.VariableSpecification] + :type variables: dict[str, ~azure.synapse.artifacts.models.VariableSpecification] :param concurrency: The max number of concurrent runs for the pipeline. :type concurrency: int :param annotations: List of tags that can be used for describing the Pipeline. @@ -28408,7 +28749,7 @@ class PipelineResource(SubResource): :type run_dimensions: dict[str, any] :param folder: The folder that this Pipeline is in. If not specified, Pipeline will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineFolder + :type folder: ~azure.synapse.artifacts.models.PipelineFolder """ _validation = { @@ -28481,7 +28822,7 @@ class PipelineRun(msrest.serialization.Model): run. :vartype parameters: dict[str, str] :ivar invoked_by: Entity that started the pipeline run. - :vartype invoked_by: ~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineRunInvokedBy + :vartype invoked_by: ~azure.synapse.artifacts.models.PipelineRunInvokedBy :ivar last_updated: The last updated timestamp for the pipeline run event in ISO8601 format. :vartype last_updated: ~datetime.datetime :ivar run_start: The start time of a pipeline run in ISO8601 format. @@ -28590,7 +28931,7 @@ class PipelineRunsQueryResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of pipeline runs. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineRun] + :type value: list[~azure.synapse.artifacts.models.PipelineRun] :param continuation_token: The continuation token for getting the next page of results, if any remaining results exist, null otherwise. :type continuation_token: str @@ -28624,8 +28965,7 @@ class PolybaseSettings(msrest.serialization.Model): collection. :type additional_properties: dict[str, any] :param reject_type: Reject type. Possible values include: "value", "percentage". - :type reject_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.PolybaseSettingsRejectType + :type reject_type: str or ~azure.synapse.artifacts.models.PolybaseSettingsRejectType :param reject_value: Specifies the value or the percentage of rows that can be rejected before the query fails. Type: number (or Expression with resultType number), minimum: 0. :type reject_value: any @@ -28676,20 +29016,17 @@ class PostgreSqlLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The connection string. :type connection_string: any :param password: The Azure key vault secret reference of password in connection string. - :type password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type password: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -28813,16 +29150,14 @@ class PostgreSqlTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -28887,13 +29222,11 @@ class PrestoLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The IP address or host name of the Presto server. (i.e. @@ -28908,12 +29241,11 @@ class PrestoLinkedService(LinkedService): :type port: any :param authentication_type: Required. The authentication mechanism used to connect to the Presto server. Possible values include: "Anonymous", "LDAP". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.PrestoAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.PrestoAuthenticationType :param username: The user name used to connect to the Presto server. :type username: any :param password: The password corresponding to the user name. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The default value is false. :type enable_ssl: any @@ -29031,16 +29363,14 @@ class PrestoObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -29230,11 +29560,11 @@ class PrivateEndpointConnection(ProxyResource): "Microsoft.Storage/storageAccounts". :vartype type: str :param private_endpoint: The private endpoint which the connection belongs to. - :type private_endpoint: ~azure.synapse.artifacts.v2021_06_01_preview.models.PrivateEndpoint + :type private_endpoint: ~azure.synapse.artifacts.models.PrivateEndpoint :param private_link_service_connection_state: Connection state of the private endpoint connection. :type private_link_service_connection_state: - ~azure.synapse.artifacts.v2021_06_01_preview.models.PrivateLinkServiceConnectionState + ~azure.synapse.artifacts.models.PrivateLinkServiceConnectionState :ivar provisioning_state: Provisioning state of the private endpoint connection. :vartype provisioning_state: str """ @@ -29329,7 +29659,7 @@ class QueryDataFlowDebugSessionsResponse(msrest.serialization.Model): """A list of active debug sessions. :param value: Array with all active debug sessions. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowDebugSessionInfo] + :type value: list[~azure.synapse.artifacts.models.DataFlowDebugSessionInfo] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -29362,13 +29692,11 @@ class QuickBooksLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_properties: Properties used to connect to QuickBooks. It is mutually @@ -29382,11 +29710,11 @@ class QuickBooksLinkedService(LinkedService): :param consumer_key: Required. The consumer key for OAuth 1.0 authentication. :type consumer_key: any :param consumer_secret: Required. The consumer secret for OAuth 1.0 authentication. - :type consumer_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type consumer_secret: ~azure.synapse.artifacts.models.SecretBase :param access_token: Required. The access token for OAuth 1.0 authentication. - :type access_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type access_token: ~azure.synapse.artifacts.models.SecretBase :param access_token_secret: Required. The access token secret for OAuth 1.0 authentication. - :type access_token_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type access_token_secret: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -29475,16 +29803,14 @@ class QuickBooksObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -29599,12 +29925,11 @@ class RecurrenceSchedule(msrest.serialization.Model): :param hours: The hours. :type hours: list[int] :param week_days: The days of the week. - :type week_days: list[str or ~azure.synapse.artifacts.v2021_06_01_preview.models.DayOfWeek] + :type week_days: list[str or ~azure.synapse.artifacts.models.DayOfWeek] :param month_days: The month days. :type month_days: list[int] :param monthly_occurrences: The monthly occurrences. - :type monthly_occurrences: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.RecurrenceScheduleOccurrence] + :type monthly_occurrences: list[~azure.synapse.artifacts.models.RecurrenceScheduleOccurrence] """ _attribute_map = { @@ -29644,7 +29969,7 @@ class RecurrenceScheduleOccurrence(msrest.serialization.Model): :type additional_properties: dict[str, any] :param day: The day of the week. Possible values include: "Sunday", "Monday", "Tuesday", "Wednesday", "Thursday", "Friday", "Saturday". - :type day: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.DayOfWeek + :type day: str or ~azure.synapse.artifacts.models.DayOfWeek :param occurrence: The occurrence. :type occurrence: int """ @@ -29718,8 +30043,7 @@ class RedshiftUnloadSettings(msrest.serialization.Model): :param s3_linked_service_name: Required. The name of the Amazon S3 linked service which will be used for the unload operation when copying from the Amazon Redshift source. - :type s3_linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type s3_linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param bucket_name: Required. The bucket of the interim Amazon S3 which will be used to store the unloaded data from Amazon Redshift source. The bucket must be in the same region as the Amazon Redshift source. Type: string (or Expression with resultType string). @@ -29824,16 +30148,14 @@ class RelationalTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The relational table name. Type: string (or Expression with resultType string). :type table_name: any @@ -29884,7 +30206,7 @@ class RerunTriggerListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of rerun triggers. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.RerunTriggerResource] + :type value: list[~azure.synapse.artifacts.models.RerunTriggerResource] :ivar next_link: The continuation token for getting the next page of results, if any remaining results exist, null otherwise. :vartype next_link: str @@ -29929,8 +30251,7 @@ class RerunTriggerResource(SubResource): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Properties of the rerun trigger. - :type properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.RerunTumblingWindowTrigger + :type properties: ~azure.synapse.artifacts.models.RerunTumblingWindowTrigger """ _validation = { @@ -29975,8 +30296,7 @@ class RerunTumblingWindowTrigger(Trigger): :type description: str :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRuntimeState + :vartype runtime_state: str or ~azure.synapse.artifacts.models.TriggerRuntimeState :param annotations: List of tags that can be used for describing the trigger. :type annotations: list[any] :param parent_trigger: Required. The parent trigger reference. @@ -30086,13 +30406,11 @@ class ResponsysLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param endpoint: Required. The endpoint of the Responsys server. @@ -30102,7 +30420,7 @@ class ResponsysLinkedService(LinkedService): :type client_id: any :param client_secret: The client secret associated with the Responsys application. Type: string (or Expression with resultType string). - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type client_secret: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. Type: boolean (or Expression with resultType boolean). :type use_encrypted_endpoints: any @@ -30189,16 +30507,14 @@ class ResponsysObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -30321,16 +30637,14 @@ class RestResourceDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param relative_url: The relative URL to the resource that the RESTful API provides. Type: string (or Expression with resultType string). :type relative_url: any @@ -30408,13 +30722,11 @@ class RestServiceLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param url: Required. The base URL of the REST service. @@ -30426,18 +30738,17 @@ class RestServiceLinkedService(LinkedService): :param authentication_type: Required. Type of authentication used to connect to the REST service. Possible values include: "Anonymous", "Basic", "AadServicePrincipal", "ManagedServiceIdentity". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.RestServiceAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.RestServiceAuthenticationType :param user_name: The user name used in Basic authentication type. :type user_name: any :param password: The password used in Basic authentication type. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param service_principal_id: The application's client ID used in AadServicePrincipal authentication type. :type service_principal_id: any :param service_principal_key: The application's key used in AadServicePrincipal authentication type. - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param tenant: The tenant information (domain name or tenant ID) used in AadServicePrincipal authentication type under which your application resides. :type tenant: any @@ -30737,9 +31048,9 @@ class RunFilterParameters(msrest.serialization.Model): 'ISO 8601' format. :type last_updated_before: ~datetime.datetime :param filters: List of filters. - :type filters: list[~azure.synapse.artifacts.v2021_06_01_preview.models.RunQueryFilter] + :type filters: list[~azure.synapse.artifacts.models.RunQueryFilter] :param order_by: List of OrderBy option. - :type order_by: list[~azure.synapse.artifacts.v2021_06_01_preview.models.RunQueryOrderBy] + :type order_by: list[~azure.synapse.artifacts.models.RunQueryOrderBy] """ _validation = { @@ -30784,11 +31095,10 @@ class RunQueryFilter(msrest.serialization.Model): runs are TriggerName, TriggerRunTimestamp and Status. Possible values include: "PipelineName", "Status", "RunStart", "RunEnd", "ActivityName", "ActivityRunStart", "ActivityRunEnd", "ActivityType", "TriggerName", "TriggerRunTimestamp", "RunGroupId", "LatestOnly". - :type operand: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.RunQueryFilterOperand + :type operand: str or ~azure.synapse.artifacts.models.RunQueryFilterOperand :param operator: Required. Operator to be used for filter. Possible values include: "Equals", "NotEquals", "In", "NotIn". - :type operator: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.RunQueryFilterOperator + :type operator: str or ~azure.synapse.artifacts.models.RunQueryFilterOperator :param values: Required. List of filter values. :type values: list[str] """ @@ -30830,9 +31140,9 @@ class RunQueryOrderBy(msrest.serialization.Model): TriggerRunTimestamp and Status. Possible values include: "RunStart", "RunEnd", "PipelineName", "Status", "ActivityName", "ActivityRunStart", "ActivityRunEnd", "TriggerName", "TriggerRunTimestamp". - :type order_by: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.RunQueryOrderByField + :type order_by: str or ~azure.synapse.artifacts.models.RunQueryOrderByField :param order: Required. Sorting order of the parameter. Possible values include: "ASC", "DESC". - :type order: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.RunQueryOrder + :type order: str or ~azure.synapse.artifacts.models.RunQueryOrder """ _validation = { @@ -30868,13 +31178,11 @@ class SalesforceLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param environment_url: The URL of Salesforce instance. Default is @@ -30886,9 +31194,9 @@ class SalesforceLinkedService(LinkedService): (or Expression with resultType string). :type username: any :param password: The password for Basic authentication of the Salesforce instance. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param security_token: The security token is optional to remotely access Salesforce instance. - :type security_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type security_token: ~azure.synapse.artifacts.models.SecretBase :param api_version: The Salesforce API version used in ADF. Type: string (or Expression with resultType string). :type api_version: any @@ -30954,13 +31262,11 @@ class SalesforceMarketingCloudLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_properties: Properties used to connect to Salesforce Marketing Cloud. It is @@ -30971,7 +31277,7 @@ class SalesforceMarketingCloudLinkedService(LinkedService): :type client_id: any :param client_secret: The client secret associated with the Salesforce Marketing Cloud application. Type: string (or Expression with resultType string). - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type client_secret: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. Type: boolean (or Expression with resultType boolean). :type use_encrypted_endpoints: any @@ -31057,16 +31363,14 @@ class SalesforceMarketingCloudObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -31189,16 +31493,14 @@ class SalesforceObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param object_api_name: The Salesforce object API name. Type: string (or Expression with resultType string). :type object_api_name: any @@ -31252,13 +31554,11 @@ class SalesforceServiceCloudLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param environment_url: The URL of Salesforce Service Cloud instance. Default is @@ -31270,9 +31570,9 @@ class SalesforceServiceCloudLinkedService(LinkedService): (or Expression with resultType string). :type username: any :param password: The password for Basic authentication of the Salesforce instance. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param security_token: The security token is optional to remotely access Salesforce instance. - :type security_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type security_token: ~azure.synapse.artifacts.models.SecretBase :param api_version: The Salesforce API version used in ADF. Type: string (or Expression with resultType string). :type api_version: any @@ -31352,16 +31652,14 @@ class SalesforceServiceCloudObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param object_api_name: The Salesforce Service Cloud object API name. Type: string (or Expression with resultType string). :type object_api_name: any @@ -31431,8 +31729,7 @@ class SalesforceServiceCloudSink(CopySink): :type max_concurrent_connections: any :param write_behavior: The write behavior for the operation. Default is Insert. Possible values include: "Insert", "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SalesforceSinkWriteBehavior + :type write_behavior: str or ~azure.synapse.artifacts.models.SalesforceSinkWriteBehavior :param external_id_field_name: The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). :type external_id_field_name: any @@ -31506,8 +31803,7 @@ class SalesforceServiceCloudSource(CopySource): :type query: any :param read_behavior: The read behavior for the operation. Default is Query. Possible values include: "Query", "QueryAll". - :type read_behavior: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SalesforceSourceReadBehavior + :type read_behavior: str or ~azure.synapse.artifacts.models.SalesforceSourceReadBehavior :param additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :type additional_columns: any @@ -31574,8 +31870,7 @@ class SalesforceSink(CopySink): :type max_concurrent_connections: any :param write_behavior: The write behavior for the operation. Default is Insert. Possible values include: "Insert", "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SalesforceSinkWriteBehavior + :type write_behavior: str or ~azure.synapse.artifacts.models.SalesforceSinkWriteBehavior :param external_id_field_name: The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). :type external_id_field_name: any @@ -31655,8 +31950,7 @@ class SalesforceSource(TabularSource): :type query: any :param read_behavior: The read behavior for the operation. Default is Query. Possible values include: "Query", "QueryAll". - :type read_behavior: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SalesforceSourceReadBehavior + :type read_behavior: str or ~azure.synapse.artifacts.models.SalesforceSourceReadBehavior """ _validation = { @@ -31713,16 +32007,14 @@ class SapBwCubeDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder """ _validation = { @@ -31770,13 +32062,11 @@ class SapBWLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param server: Required. Host name of the SAP BW instance. Type: string (or Expression with @@ -31792,7 +32082,7 @@ class SapBWLinkedService(LinkedService): resultType string). :type user_name: any :param password: Password to access the SAP BW server. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -31919,13 +32209,11 @@ class SapCloudForCustomerLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param url: Required. The URL of SAP Cloud for Customer OData API. For example, @@ -31936,7 +32224,7 @@ class SapCloudForCustomerLinkedService(LinkedService): resultType string). :type username: any :param password: The password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Either encryptedCredential or username/password must be provided. Type: string (or Expression with resultType string). @@ -32002,16 +32290,14 @@ class SapCloudForCustomerResourceDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param path: Required. The path of the SAP Cloud for Customer OData entity. Type: string (or Expression with resultType string). :type path: any @@ -32083,7 +32369,7 @@ class SapCloudForCustomerSink(CopySink): :param write_behavior: The write behavior for the operation. Default is 'Insert'. Possible values include: "Insert", "Update". :type write_behavior: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SapCloudForCustomerSinkWriteBehavior + ~azure.synapse.artifacts.models.SapCloudForCustomerSinkWriteBehavior :param http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string (or Expression with resultType string), pattern: @@ -32207,13 +32493,11 @@ class SapEccLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param url: Required. The URL of SAP ECC OData API. For example, @@ -32224,7 +32508,7 @@ class SapEccLinkedService(LinkedService): resultType string). :type username: str :param password: The password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Either encryptedCredential or username/password must be provided. Type: string (or Expression with resultType string). @@ -32290,16 +32574,14 @@ class SapEccResourceDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param path: Required. The path of the SAP ECC OData entity. Type: string (or Expression with resultType string). :type path: any @@ -32424,13 +32706,11 @@ class SapHanaLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: SAP HANA ODBC connection string. Type: string, SecureString or @@ -32441,13 +32721,12 @@ class SapHanaLinkedService(LinkedService): :type server: any :param authentication_type: The authentication type to be used to connect to the SAP HANA server. Possible values include: "Basic", "Windows". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SapHanaAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.SapHanaAuthenticationType :param user_name: Username to access the SAP HANA server. Type: string (or Expression with resultType string). :type user_name: any :param password: Password to access the SAP HANA server. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -32554,12 +32833,10 @@ class SapHanaSource(TabularSource): :type packet_size: any :param partition_option: The partition mechanism that will be used for SAP HANA read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "SapHanaDynamicRange". - :type partition_option: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SapHanaPartitionOption + :type partition_option: str or ~azure.synapse.artifacts.models.SapHanaPartitionOption :param partition_settings: The settings that will be leveraged for SAP HANA source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SapHanaPartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.SapHanaPartitionSettings """ _validation = { @@ -32622,16 +32899,14 @@ class SapHanaTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param schema_type_properties_schema: The schema name of SAP HANA. Type: string (or Expression with resultType string). :type schema_type_properties_schema: any @@ -32690,13 +32965,11 @@ class SapOpenHubLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param server: Required. Host name of the SAP BW instance where the open hub destination is @@ -32721,7 +32994,7 @@ class SapOpenHubLinkedService(LinkedService): :type user_name: any :param password: Password to access the SAP BW server where the open hub destination is located. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param message_server: The hostname of the SAP Message Server. Type: string (or Expression with resultType string). :type message_server: any @@ -32901,16 +33174,14 @@ class SapOpenHubTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param open_hub_destination_name: Required. The name of the Open Hub Destination with destination type as Database Table. Type: string (or Expression with resultType string). :type open_hub_destination_name: any @@ -32978,13 +33249,11 @@ class SapTableLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param server: Host name of the SAP instance where the table is located. Type: string (or @@ -33008,7 +33277,7 @@ class SapTableLinkedService(LinkedService): (or Expression with resultType string). :type user_name: any :param password: Password to access the SAP server where the table is located. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param message_server: The hostname of the SAP Message Server. Type: string (or Expression with resultType string). :type message_server: any @@ -33175,16 +33444,14 @@ class SapTableResourceDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: Required. The name of the SAP Table. Type: string (or Expression with resultType string). :type table_name: any @@ -33278,12 +33545,10 @@ class SapTableSource(TabularSource): :param partition_option: The partition mechanism that will be used for SAP table read in parallel. Possible values include: "None", "PartitionOnInt", "PartitionOnCalendarYear", "PartitionOnCalendarMonth", "PartitionOnCalendarDate", "PartitionOnTime". - :type partition_option: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SapTablePartitionOption + :type partition_option: str or ~azure.synapse.artifacts.models.SapTablePartitionOption :param partition_settings: The settings that will be leveraged for SAP table source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SapTablePartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.SapTablePartitionSettings """ _validation = { @@ -33358,15 +33623,13 @@ class ScheduleTrigger(MultiplePipelineTrigger): :type description: str :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRuntimeState + :vartype runtime_state: str or ~azure.synapse.artifacts.models.TriggerRuntimeState :param annotations: List of tags that can be used for describing the trigger. :type annotations: list[any] :param pipelines: Pipelines that need to be started. - :type pipelines: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerPipelineReference] + :type pipelines: list[~azure.synapse.artifacts.models.TriggerPipelineReference] :param recurrence: Required. Recurrence schedule configuration. - :type recurrence: ~azure.synapse.artifacts.v2021_06_01_preview.models.ScheduleTriggerRecurrence + :type recurrence: ~azure.synapse.artifacts.models.ScheduleTriggerRecurrence """ _validation = { @@ -33408,7 +33671,7 @@ class ScheduleTriggerRecurrence(msrest.serialization.Model): :type additional_properties: dict[str, any] :param frequency: The frequency. Possible values include: "NotSpecified", "Minute", "Hour", "Day", "Week", "Month", "Year". - :type frequency: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.RecurrenceFrequency + :type frequency: str or ~azure.synapse.artifacts.models.RecurrenceFrequency :param interval: The interval. :type interval: int :param start_time: The start time. @@ -33418,7 +33681,7 @@ class ScheduleTriggerRecurrence(msrest.serialization.Model): :param time_zone: The time zone. :type time_zone: str :param schedule: The recurrence schedule. - :type schedule: ~azure.synapse.artifacts.v2021_06_01_preview.models.RecurrenceSchedule + :type schedule: ~azure.synapse.artifacts.models.RecurrenceSchedule """ _attribute_map = { @@ -33462,9 +33725,8 @@ class ScriptAction(msrest.serialization.Model): :type name: str :param uri: Required. The URI for the script action. :type uri: str - :param roles: Required. The node types on which the script action should be executed. Possible - values include: "Headnode", "Workernode", "Zookeeper". - :type roles: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.HdiNodeTypes + :param roles: Required. The node types on which the script action should be executed. + :type roles: any :param parameters: The parameters for the script action. :type parameters: str """ @@ -33478,7 +33740,7 @@ class ScriptAction(msrest.serialization.Model): _attribute_map = { 'name': {'key': 'name', 'type': 'str'}, 'uri': {'key': 'uri', 'type': 'str'}, - 'roles': {'key': 'roles', 'type': 'str'}, + 'roles': {'key': 'roles', 'type': 'object'}, 'parameters': {'key': 'parameters', 'type': 'str'}, } @@ -33487,7 +33749,7 @@ def __init__( *, name: str, uri: str, - roles: Union[str, "HdiNodeTypes"], + roles: Any, parameters: Optional[str] = None, **kwargs ): @@ -33580,12 +33842,11 @@ class SelfHostedIntegrationRuntime(IntegrationRuntime): :type additional_properties: dict[str, any] :param type: Required. Type of integration runtime.Constant filled by server. Possible values include: "Managed", "SelfHosted". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeType + :type type: str or ~azure.synapse.artifacts.models.IntegrationRuntimeType :param description: Integration runtime description. :type description: str :param linked_info: Linked integration runtime type from data factory. - :type linked_info: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedIntegrationRuntimeType + :type linked_info: ~azure.synapse.artifacts.models.LinkedIntegrationRuntimeType """ _validation = { @@ -33623,13 +33884,11 @@ class ServiceNowLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param endpoint: Required. The endpoint of the ServiceNow server. (i.e. @@ -33637,18 +33896,17 @@ class ServiceNowLinkedService(LinkedService): :type endpoint: any :param authentication_type: Required. The authentication type to use. Possible values include: "Basic", "OAuth2". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.ServiceNowAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.ServiceNowAuthenticationType :param username: The user name used to connect to the ServiceNow server for Basic and OAuth2 authentication. :type username: any :param password: The password corresponding to the user name for Basic and OAuth2 authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param client_id: The client id for OAuth2 authentication. :type client_id: any :param client_secret: The client secret for OAuth2 authentication. - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type client_secret: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -33743,16 +34001,14 @@ class ServiceNowObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -33871,9 +34127,9 @@ class SetVariableActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param variable_name: Name of the variable whose value needs to be set. :type variable_name: str :param value: Value to be set. Could be a static value or Expression. @@ -34056,13 +34312,11 @@ class SftpServerLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The SFTP server host name. Type: string (or Expression with resultType @@ -34073,13 +34327,12 @@ class SftpServerLinkedService(LinkedService): :type port: any :param authentication_type: The authentication type to be used to connect to the FTP server. Possible values include: "Basic", "SshPublicKey". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SftpAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.SftpAuthenticationType :param user_name: The username used to log on to the SFTP server. Type: string (or Expression with resultType string). :type user_name: any :param password: Password to logon the SFTP server for Basic authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -34092,10 +34345,10 @@ class SftpServerLinkedService(LinkedService): :param private_key_content: Base64 encoded SSH private key content for SshPublicKey authentication. For on-premises copy with SshPublicKey authentication, either PrivateKeyPath or PrivateKeyContent should be specified. SSH private key should be OpenSSH format. - :type private_key_content: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type private_key_content: ~azure.synapse.artifacts.models.SecretBase :param pass_phrase: The password to decrypt the SSH private key if the SSH private key is encrypted. - :type pass_phrase: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type pass_phrase: ~azure.synapse.artifacts.models.SecretBase :param skip_host_key_validation: If true, skip the SSH host key validation. Default value is false. Type: boolean (or Expression with resultType boolean). :type skip_host_key_validation: any @@ -34230,13 +34483,11 @@ class SharePointOnlineListLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param site_url: Required. The URL of the SharePoint Online site. For example, @@ -34253,7 +34504,7 @@ class SharePointOnlineListLinkedService(LinkedService): :type service_principal_id: any :param service_principal_key: Required. The client secret of your application registered in Azure Active Directory. Type: string (or Expression with resultType string). - :type service_principal_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type service_principal_key: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -34325,16 +34576,14 @@ class SharePointOnlineListResourceDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param list_name: The name of the SharePoint Online list. Type: string (or Expression with resultType string). :type list_name: any @@ -34447,20 +34696,18 @@ class ShopifyLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. The endpoint of the Shopify server. (i.e. mystore.myshopify.com). :type host: any :param access_token: The API access token that can be used to access Shopify’s data. The token won't expire if it is offline mode. - :type access_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type access_token: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -34542,16 +34789,14 @@ class ShopifyObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -34734,16 +34979,14 @@ class SnowflakeDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param schema_type_properties_schema: The schema name of the Snowflake database. Type: string (or Expression with resultType string). :type schema_type_properties_schema: any @@ -34897,21 +35140,18 @@ class SnowflakeLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The connection string of snowflake. Type: string, SecureString. :type connection_string: any :param password: The Azure key vault secret reference of password in connection string. - :type password: - ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type password: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -34984,8 +35224,7 @@ class SnowflakeSink(CopySink): string). :type pre_copy_script: any :param import_settings: Snowflake import settings. - :type import_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SnowflakeImportCopyCommand + :type import_settings: ~azure.synapse.artifacts.models.SnowflakeImportCopyCommand """ _validation = { @@ -35045,8 +35284,7 @@ class SnowflakeSource(CopySource): :param query: Snowflake Sql query. Type: string (or Expression with resultType string). :type query: any :param export_settings: Snowflake export settings. - :type export_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SnowflakeExportCopyCommand + :type export_settings: ~azure.synapse.artifacts.models.SnowflakeExportCopyCommand """ _validation = { @@ -35086,7 +35324,7 @@ class SparkBatchJob(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param livy_info: - :type livy_info: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkBatchJobState + :type livy_info: ~azure.synapse.artifacts.models.SparkBatchJobState :param name: The batch name. :type name: str :param workspace_name: The workspace name. @@ -35100,17 +35338,16 @@ class SparkBatchJob(msrest.serialization.Model): :param artifact_id: The artifact identifier. :type artifact_id: str :param job_type: The job type. Possible values include: "SparkBatch", "SparkSession". - :type job_type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobType + :type job_type: str or ~azure.synapse.artifacts.models.SparkJobType :param result: The Spark batch job result. Possible values include: "Uncertain", "Succeeded", "Failed", "Cancelled". - :type result: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkBatchJobResultType + :type result: str or ~azure.synapse.artifacts.models.SparkBatchJobResultType :param scheduler: The scheduler information. - :type scheduler: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkScheduler + :type scheduler: ~azure.synapse.artifacts.models.SparkScheduler :param plugin: The plugin information. - :type plugin: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkServicePlugin + :type plugin: ~azure.synapse.artifacts.models.SparkServicePlugin :param errors: The error information. - :type errors: list[~azure.synapse.artifacts.v2021_06_01_preview.models.SparkServiceError] + :type errors: list[~azure.synapse.artifacts.models.SparkServiceError] :param tags: A set of tags. The tags. :type tags: dict[str, str] :param id: Required. The session Id. @@ -35119,8 +35356,9 @@ class SparkBatchJob(msrest.serialization.Model): :type app_id: str :param app_info: The detailed application info. :type app_info: dict[str, str] - :param state: The batch state. - :type state: str + :param state: The batch state. Possible values include: "not_started", "starting", "idle", + "busy", "shutting_down", "error", "dead", "killed", "success", "running", "recovering". + :type state: str or ~azure.synapse.artifacts.models.LivyStates :param log_lines: The log lines. :type log_lines: list[str] """ @@ -35169,7 +35407,7 @@ def __init__( tags: Optional[Dict[str, str]] = None, app_id: Optional[str] = None, app_info: Optional[Dict[str, str]] = None, - state: Optional[str] = None, + state: Optional[Union[str, "LivyStates"]] = None, log_lines: Optional[List[str]] = None, **kwargs ): @@ -35214,7 +35452,7 @@ class SparkBatchJobState(msrest.serialization.Model): :param current_state: the Spark job state. :type current_state: str :param job_creation_request: - :type job_creation_request: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkRequest + :type job_creation_request: ~azure.synapse.artifacts.models.SparkRequest """ _attribute_map = { @@ -35272,6 +35510,8 @@ class SparkConfiguration(msrest.serialization.Model): :type created_by: str :param created: The timestamp of resource creation. :type created: ~datetime.datetime + :param config_merge_rule: SparkConfiguration configMergeRule. + :type config_merge_rule: dict[str, str] """ _validation = { @@ -35285,6 +35525,7 @@ class SparkConfiguration(msrest.serialization.Model): 'notes': {'key': 'notes', 'type': 'str'}, 'created_by': {'key': 'createdBy', 'type': 'str'}, 'created': {'key': 'created', 'type': 'iso-8601'}, + 'config_merge_rule': {'key': 'configMergeRule', 'type': '{str}'}, } def __init__( @@ -35296,6 +35537,7 @@ def __init__( notes: Optional[str] = None, created_by: Optional[str] = None, created: Optional[datetime.datetime] = None, + config_merge_rule: Optional[Dict[str, str]] = None, **kwargs ): super(SparkConfiguration, self).__init__(**kwargs) @@ -35305,6 +35547,7 @@ def __init__( self.notes = notes self.created_by = created_by self.created = created + self.config_merge_rule = config_merge_rule class SparkConfigurationListResponse(msrest.serialization.Model): @@ -35313,8 +35556,7 @@ class SparkConfigurationListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of sparkconfigurations. - :type value: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.SparkConfigurationResource] + :type value: list[~azure.synapse.artifacts.models.SparkConfigurationResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -35358,7 +35600,7 @@ class SparkConfigurationResource(SubResource): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Properties of Spark Configuration. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkConfiguration + :type properties: ~azure.synapse.artifacts.models.SparkConfiguration """ _validation = { @@ -35398,14 +35640,16 @@ class SparkJobDefinition(msrest.serialization.Model): :param description: The description of the Spark job definition. :type description: str :param target_big_data_pool: Required. Big data pool reference. - :type target_big_data_pool: - ~azure.synapse.artifacts.v2021_06_01_preview.models.BigDataPoolReference + :type target_big_data_pool: ~azure.synapse.artifacts.models.BigDataPoolReference :param required_spark_version: The required Spark version of the application. :type required_spark_version: str :param language: The language of the Spark application. :type language: str :param job_properties: Required. The properties of the Spark job. - :type job_properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobProperties + :type job_properties: ~azure.synapse.artifacts.models.SparkJobProperties + :param folder: The folder that this Spark job definition is in. If not specified, this Spark + job definition will appear at the root level. + :type folder: ~azure.synapse.artifacts.models.SparkJobDefinitionFolder """ _validation = { @@ -35420,6 +35664,7 @@ class SparkJobDefinition(msrest.serialization.Model): 'required_spark_version': {'key': 'requiredSparkVersion', 'type': 'str'}, 'language': {'key': 'language', 'type': 'str'}, 'job_properties': {'key': 'jobProperties', 'type': 'SparkJobProperties'}, + 'folder': {'key': 'folder', 'type': 'SparkJobDefinitionFolder'}, } def __init__( @@ -35431,6 +35676,7 @@ def __init__( description: Optional[str] = None, required_spark_version: Optional[str] = None, language: Optional[str] = None, + folder: Optional["SparkJobDefinitionFolder"] = None, **kwargs ): super(SparkJobDefinition, self).__init__(**kwargs) @@ -35440,6 +35686,28 @@ def __init__( self.required_spark_version = required_spark_version self.language = language self.job_properties = job_properties + self.folder = folder + + +class SparkJobDefinitionFolder(msrest.serialization.Model): + """The folder that this Spark job definition is in. If not specified, this Spark job definition will appear at the root level. + + :param name: The name of the folder that this Spark job definition is in. + :type name: str + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + } + + def __init__( + self, + *, + name: Optional[str] = None, + **kwargs + ): + super(SparkJobDefinitionFolder, self).__init__(**kwargs) + self.name = name class SparkJobDefinitionResource(SubResource): @@ -35460,7 +35728,7 @@ class SparkJobDefinitionResource(SubResource): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Properties of spark job definition. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobDefinition + :type properties: ~azure.synapse.artifacts.models.SparkJobDefinition """ _validation = { @@ -35495,8 +35763,7 @@ class SparkJobDefinitionsListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of spark job definitions. - :type value: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobDefinitionResource] + :type value: list[~azure.synapse.artifacts.models.SparkJobDefinitionResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -35631,13 +35898,11 @@ class SparkLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param host: Required. IP address or host name of the Spark server. @@ -35647,21 +35912,20 @@ class SparkLinkedService(LinkedService): :type port: any :param server_type: The type of Spark server. Possible values include: "SharkServer", "SharkServer2", "SparkThriftServer". - :type server_type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkServerType + :type server_type: str or ~azure.synapse.artifacts.models.SparkServerType :param thrift_transport_protocol: The transport protocol to use in the Thrift layer. Possible values include: "Binary", "SASL", "HTTP ". :type thrift_transport_protocol: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkThriftTransportProtocol + ~azure.synapse.artifacts.models.SparkThriftTransportProtocol :param authentication_type: Required. The authentication method used to access the Spark server. Possible values include: "Anonymous", "Username", "UsernameAndPassword", "WindowsAzureHDInsightService". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.SparkAuthenticationType :param username: The user name that you use to access Spark Server. :type username: any :param password: The password corresponding to the user name that you provided in the Username field. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param http_path: The partial URL corresponding to the Spark server. :type http_path: any :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The @@ -35777,16 +36041,14 @@ class SparkObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -35938,8 +36200,7 @@ class SparkScheduler(msrest.serialization.Model): :param cancellation_requested_at: :type cancellation_requested_at: ~datetime.datetime :param current_state: Possible values include: "Queued", "Scheduled", "Ended". - :type current_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SchedulerCurrentState + :type current_state: str or ~azure.synapse.artifacts.models.SchedulerCurrentState """ _attribute_map = { @@ -35976,7 +36237,7 @@ class SparkServiceError(msrest.serialization.Model): :param error_code: :type error_code: str :param source: Possible values include: "System", "User", "Unknown", "Dependency". - :type source: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkErrorSource + :type source: str or ~azure.synapse.artifacts.models.SparkErrorSource """ _attribute_map = { @@ -36014,8 +36275,7 @@ class SparkServicePlugin(msrest.serialization.Model): :type cleanup_started_at: ~datetime.datetime :param current_state: Possible values include: "Preparation", "ResourceAcquisition", "Queued", "Submission", "Monitoring", "Cleanup", "Ended". - :type current_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.PluginCurrentState + :type current_state: str or ~azure.synapse.artifacts.models.PluginCurrentState """ _attribute_map = { @@ -36112,41 +36372,43 @@ def __init__( class SqlConnection(msrest.serialization.Model): """The connection used to execute the SQL script. - All required parameters must be populated in order to send to Azure. - :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. The type of the connection. Possible values include: "SqlOnDemand", - "SqlPool". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlConnectionType - :param name: Required. The identifier of the connection. + :param type: The type of the connection. Possible values include: "SqlOnDemand", "SqlPool". + :type type: str or ~azure.synapse.artifacts.models.SqlConnectionType + :param name: The identifier of the connection. :type name: str + :param pool_name: The associated SQL pool name (supported by SQL pool v3). + :type pool_name: str + :param database_name: The associated database name (supported by SQL pool v3). + :type database_name: str """ - _validation = { - 'type': {'required': True}, - 'name': {'required': True}, - } - _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, 'type': {'key': 'type', 'type': 'str'}, 'name': {'key': 'name', 'type': 'str'}, + 'pool_name': {'key': 'poolName', 'type': 'str'}, + 'database_name': {'key': 'databaseName', 'type': 'str'}, } def __init__( self, *, - type: Union[str, "SqlConnectionType"], - name: str, additional_properties: Optional[Dict[str, Any]] = None, + type: Optional[Union[str, "SqlConnectionType"]] = None, + name: Optional[str] = None, + pool_name: Optional[str] = None, + database_name: Optional[str] = None, **kwargs ): super(SqlConnection, self).__init__(**kwargs) self.additional_properties = additional_properties self.type = type self.name = name + self.pool_name = pool_name + self.database_name = database_name class SqlDWSink(CopySink): @@ -36181,14 +36443,13 @@ class SqlDWSink(CopySink): applicable. Type: boolean (or Expression with resultType boolean). :type allow_poly_base: any :param poly_base_settings: Specifies PolyBase-related settings when allowPolyBase is true. - :type poly_base_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.PolybaseSettings + :type poly_base_settings: ~azure.synapse.artifacts.models.PolybaseSettings :param allow_copy_command: Indicates to use Copy Command to copy data into SQL Data Warehouse. Type: boolean (or Expression with resultType boolean). :type allow_copy_command: any :param copy_command_settings: Specifies Copy Command related settings when allowCopyCommand is true. - :type copy_command_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.DWCopyCommandSettings + :type copy_command_settings: ~azure.synapse.artifacts.models.DWCopyCommandSettings :param table_option: The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). :type table_option: any @@ -36281,8 +36542,7 @@ class SqlDWSource(TabularSource): Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". :type partition_option: any :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.SqlPartitionSettings """ _validation = { @@ -36365,7 +36625,7 @@ class SqlMISink(CopySink): :type pre_copy_script: any :param stored_procedure_parameters: SQL stored procedure parameters. :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] :param stored_procedure_table_type_parameter_name: The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). :type stored_procedure_table_type_parameter_name: any @@ -36455,15 +36715,14 @@ class SqlMISource(TabularSource): :param stored_procedure_parameters: Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] :param produce_additional_types: Which additional types to produce. :type produce_additional_types: any :param partition_option: The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". :type partition_option: any :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.SqlPartitionSettings """ _validation = { @@ -36573,7 +36832,7 @@ class SqlPool(TrackedResource): :param location: Required. The geo-location where the resource lives. :type location: str :param sku: SQL pool SKU. - :type sku: ~azure.synapse.artifacts.v2021_06_01_preview.models.Sku + :type sku: ~azure.synapse.artifacts.models.Sku :param max_size_bytes: Maximum size in bytes. :type max_size_bytes: long :param collation: Collation mode. @@ -36655,7 +36914,7 @@ class SqlPoolInfoListResult(msrest.serialization.Model): :param next_link: Link to the next page of results. :type next_link: str :param value: List of SQL pools. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPool] + :type value: list[~azure.synapse.artifacts.models.SqlPool] """ _attribute_map = { @@ -36681,7 +36940,7 @@ class SqlPoolReference(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param type: Required. SQL pool reference type. Possible values include: "SqlPoolReference". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPoolReferenceType + :type type: str or ~azure.synapse.artifacts.models.SqlPoolReferenceType :param reference_name: Required. Reference SQL pool name. :type reference_name: str """ @@ -36723,18 +36982,18 @@ class SqlPoolStoredProcedureActivity(Activity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param sql_pool: Required. SQL pool stored procedure reference. - :type sql_pool: ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPoolReference + :type sql_pool: ~azure.synapse.artifacts.models.SqlPoolReference :param stored_procedure_name: Required. Stored procedure name. Type: string (or Expression with resultType string). :type stored_procedure_name: any :param stored_procedure_parameters: Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] """ _validation = { @@ -36787,9 +37046,12 @@ class SqlScript(msrest.serialization.Model): :param description: The description of the SQL script. :type description: str :param type: The type of the SQL script. Possible values include: "SqlQuery". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlScriptType + :type type: str or ~azure.synapse.artifacts.models.SqlScriptType :param content: Required. The content of the SQL script. - :type content: ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlScriptContent + :type content: ~azure.synapse.artifacts.models.SqlScriptContent + :param folder: The folder that this SQL script is in. If not specified, this SQL script will + appear at the root level. + :type folder: ~azure.synapse.artifacts.models.SqlScriptFolder """ _validation = { @@ -36801,6 +37063,7 @@ class SqlScript(msrest.serialization.Model): 'description': {'key': 'description', 'type': 'str'}, 'type': {'key': 'type', 'type': 'str'}, 'content': {'key': 'content', 'type': 'SqlScriptContent'}, + 'folder': {'key': 'folder', 'type': 'SqlScriptFolder'}, } def __init__( @@ -36810,6 +37073,7 @@ def __init__( additional_properties: Optional[Dict[str, Any]] = None, description: Optional[str] = None, type: Optional[Union[str, "SqlScriptType"]] = None, + folder: Optional["SqlScriptFolder"] = None, **kwargs ): super(SqlScript, self).__init__(**kwargs) @@ -36817,6 +37081,7 @@ def __init__( self.description = description self.type = type self.content = content + self.folder = folder class SqlScriptContent(msrest.serialization.Model): @@ -36829,21 +37094,23 @@ class SqlScriptContent(msrest.serialization.Model): :type additional_properties: dict[str, any] :param query: Required. SQL query to execute. :type query: str - :param current_connection: Required. The connection used to execute the SQL script. - :type current_connection: ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlConnection + :param current_connection: The connection used to execute the SQL script. + :type current_connection: ~azure.synapse.artifacts.models.SqlConnection + :param result_limit: Limit of results, '-1' for no limit. + :type result_limit: int :param metadata: The metadata of the SQL script. - :type metadata: ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlScriptMetadata + :type metadata: ~azure.synapse.artifacts.models.SqlScriptMetadata """ _validation = { 'query': {'required': True}, - 'current_connection': {'required': True}, } _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, 'query': {'key': 'query', 'type': 'str'}, 'current_connection': {'key': 'currentConnection', 'type': 'SqlConnection'}, + 'result_limit': {'key': 'resultLimit', 'type': 'int'}, 'metadata': {'key': 'metadata', 'type': 'SqlScriptMetadata'}, } @@ -36851,8 +37118,9 @@ def __init__( self, *, query: str, - current_connection: "SqlConnection", additional_properties: Optional[Dict[str, Any]] = None, + current_connection: Optional["SqlConnection"] = None, + result_limit: Optional[int] = None, metadata: Optional["SqlScriptMetadata"] = None, **kwargs ): @@ -36860,9 +37128,31 @@ def __init__( self.additional_properties = additional_properties self.query = query self.current_connection = current_connection + self.result_limit = result_limit self.metadata = metadata +class SqlScriptFolder(msrest.serialization.Model): + """The folder that this SQL script is in. If not specified, this SQL script will appear at the root level. + + :param name: The name of the folder that this SQL script is in. + :type name: str + """ + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + } + + def __init__( + self, + *, + name: Optional[str] = None, + **kwargs + ): + super(SqlScriptFolder, self).__init__(**kwargs) + self.name = name + + class SqlScriptMetadata(msrest.serialization.Model): """The metadata of the SQL script. @@ -36908,7 +37198,7 @@ class SqlScriptResource(msrest.serialization.Model): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Properties of sql script. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlScript + :type properties: ~azure.synapse.artifacts.models.SqlScript """ _validation = { @@ -36948,7 +37238,7 @@ class SqlScriptsListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of sql scripts. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.SqlScriptResource] + :type value: list[~azure.synapse.artifacts.models.SqlScriptResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -36985,13 +37275,11 @@ class SqlServerLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Required. The connection string. Type: string, SecureString or @@ -37001,7 +37289,7 @@ class SqlServerLinkedService(LinkedService): with resultType string). :type user_name: any :param password: The on-premises Windows authentication password. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -37084,7 +37372,7 @@ class SqlServerSink(CopySink): :type pre_copy_script: any :param stored_procedure_parameters: SQL stored procedure parameters. :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] :param stored_procedure_table_type_parameter_name: The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). :type stored_procedure_table_type_parameter_name: any @@ -37174,15 +37462,14 @@ class SqlServerSource(TabularSource): :param stored_procedure_parameters: Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] :param produce_additional_types: Which additional types to produce. :type produce_additional_types: any :param partition_option: The partition mechanism that will be used for Sql read in parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". :type partition_option: any :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.SqlPartitionSettings """ _validation = { @@ -37247,21 +37534,20 @@ class SqlServerStoredProcedureActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param stored_procedure_name: Required. Stored procedure name. Type: string (or Expression with resultType string). :type stored_procedure_name: any :param stored_procedure_parameters: Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] """ _validation = { @@ -37322,16 +37608,14 @@ class SqlServerTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -37422,7 +37706,7 @@ class SqlSink(CopySink): :type pre_copy_script: any :param stored_procedure_parameters: SQL stored procedure parameters. :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] :param stored_procedure_table_type_parameter_name: The stored procedure parameter name of the table type. Type: string (or Expression with resultType string). :type stored_procedure_table_type_parameter_name: any @@ -37512,7 +37796,7 @@ class SqlSource(TabularSource): :param stored_procedure_parameters: Value and type setting for stored procedure parameters. Example: "{Parameter1: {value: "1", type: "int"}}". :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameter] + ~azure.synapse.artifacts.models.StoredProcedureParameter] :param isolation_level: Specifies the transaction locking behavior for the SQL source. Allowed values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is ReadCommitted. Type: string (or Expression with resultType string). @@ -37521,8 +37805,7 @@ class SqlSource(TabularSource): Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". :type partition_option: any :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.SqlPartitionSettings """ _validation = { @@ -37583,13 +37866,11 @@ class SquareLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_properties: Properties used to connect to Square. It is mutually exclusive @@ -37600,7 +37881,7 @@ class SquareLinkedService(LinkedService): :param client_id: Required. The client ID associated with your Square application. :type client_id: any :param client_secret: The client secret associated with your Square application. - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type client_secret: ~azure.synapse.artifacts.models.SecretBase :param redirect_uri: Required. The redirect URL assigned in the Square application dashboard. (i.e. http://localhost:2500). :type redirect_uri: any @@ -37696,16 +37977,14 @@ class SquareObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -37819,7 +38098,7 @@ class SSISAccessCredential(msrest.serialization.Model): :param user_name: Required. UseName for windows authentication. :type user_name: any :param password: Required. Password for windows authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase """ _validation = { @@ -37903,7 +38182,7 @@ class SSISExecutionCredential(msrest.serialization.Model): :param user_name: Required. UseName for windows authentication. :type user_name: any :param password: Required. Password for windows authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecureString + :type password: ~azure.synapse.artifacts.models.SecureString """ _validation = { @@ -37969,10 +38248,9 @@ class SSISLogLocation(msrest.serialization.Model): with resultType string). :type log_path: any :param type: Required. The type of SSIS log location. Possible values include: "File". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.SsisLogLocationType + :type type: str or ~azure.synapse.artifacts.models.SsisLogLocationType :param access_credential: The package execution log access credential. - :type access_credential: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISAccessCredential + :type access_credential: ~azure.synapse.artifacts.models.SSISAccessCredential :param log_refresh_interval: Specifies the interval to refresh log. The default interval is 5 minutes. Type: string (or Expression with resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). @@ -38051,18 +38329,16 @@ class SSISPackageLocation(msrest.serialization.Model): :type package_path: any :param type: The type of SSIS package location. Possible values include: "SSISDB", "File", "InlinePackage", "PackageStore". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.SsisPackageLocationType + :type type: str or ~azure.synapse.artifacts.models.SsisPackageLocationType :param package_password: Password of the package. - :type package_password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type package_password: ~azure.synapse.artifacts.models.SecretBase :param access_credential: The package access credential. - :type access_credential: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISAccessCredential + :type access_credential: ~azure.synapse.artifacts.models.SSISAccessCredential :param configuration_path: The configuration file of the package execution. Type: string (or Expression with resultType string). :type configuration_path: any :param configuration_access_credential: The configuration file access credential. - :type configuration_access_credential: - ~azure.synapse.artifacts.v2021_06_01_preview.models.SSISAccessCredential + :type configuration_access_credential: ~azure.synapse.artifacts.models.SSISAccessCredential :param package_name: The package name. :type package_name: str :param package_content: The embedded package content. Type: string (or Expression with @@ -38071,8 +38347,7 @@ class SSISPackageLocation(msrest.serialization.Model): :param package_last_modified_date: The embedded package last modified date. :type package_last_modified_date: str :param child_packages: The embedded child package list. - :type child_packages: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.SSISChildPackage] + :type child_packages: list[~azure.synapse.artifacts.models.SSISChildPackage] """ _attribute_map = { @@ -38159,8 +38434,7 @@ class StagingSettings(msrest.serialization.Model): collection. :type additional_properties: dict[str, any] :param linked_service_name: Required. Staging linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param path: The path to storage for storing the interim data. Type: string (or Expression with resultType string). :type path: any @@ -38202,12 +38476,11 @@ class StartDataFlowDebugSessionRequest(msrest.serialization.Model): :param session_id: The ID of data flow debug session. :type session_id: str :param data_flow: Data flow instance. - :type data_flow: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowResource + :type data_flow: ~azure.synapse.artifacts.models.DataFlowResource :param datasets: List of datasets. - :type datasets: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetResource] + :type datasets: list[~azure.synapse.artifacts.models.DatasetResource] :param linked_services: List of linked services. - :type linked_services: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceResource] + :type linked_services: list[~azure.synapse.artifacts.models.LinkedServiceResource] :param staging: Staging info for debug session. :type staging: any :param debug_settings: Data flow debug settings. @@ -38277,8 +38550,7 @@ class StoredProcedureParameter(msrest.serialization.Model): :type value: any :param type: Stored procedure parameter type. Possible values include: "String", "Int", "Int64", "Decimal", "Guid", "Boolean", "Date". - :type type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.StoredProcedureParameterType + :type type: str or ~azure.synapse.artifacts.models.StoredProcedureParameterType """ _attribute_map = { @@ -38313,19 +38585,19 @@ class SwitchActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param on: Required. An expression that would evaluate to a string or integer. This is used to determine the block of activities in cases that will be executed. - :type on: ~azure.synapse.artifacts.v2021_06_01_preview.models.Expression + :type on: ~azure.synapse.artifacts.models.Expression :param cases: List of cases that correspond to expected values of the 'on' property. This is an optional property and if not provided, the activity will execute activities provided in defaultActivities. - :type cases: list[~azure.synapse.artifacts.v2021_06_01_preview.models.SwitchCase] + :type cases: list[~azure.synapse.artifacts.models.SwitchCase] :param default_activities: List of activities to execute if no case condition is satisfied. This is an optional property and if not provided, the activity will exit without any action. - :type default_activities: list[~azure.synapse.artifacts.v2021_06_01_preview.models.Activity] + :type default_activities: list[~azure.synapse.artifacts.models.Activity] """ _validation = { @@ -38372,7 +38644,7 @@ class SwitchCase(msrest.serialization.Model): :param value: Expected value that satisfies the expression result of the 'on' property. :type value: str :param activities: List of activities to execute for satisfied case condition. - :type activities: list[~azure.synapse.artifacts.v2021_06_01_preview.models.Activity] + :type activities: list[~azure.synapse.artifacts.models.Activity] """ _attribute_map = { @@ -38403,13 +38675,11 @@ class SybaseLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param server: Required. Server name for connection. Type: string (or Expression with @@ -38422,13 +38692,12 @@ class SybaseLinkedService(LinkedService): :type schema: any :param authentication_type: AuthenticationType to be used for connection. Possible values include: "Basic", "Windows". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.SybaseAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.SybaseAuthenticationType :param username: Username for authentication. Type: string (or Expression with resultType string). :type username: any :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -38565,16 +38834,14 @@ class SybaseTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The Sybase table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -38631,16 +38898,15 @@ class SynapseNotebookActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param notebook: Required. Synapse notebook reference. - :type notebook: ~azure.synapse.artifacts.v2021_06_01_preview.models.SynapseNotebookReference + :type notebook: ~azure.synapse.artifacts.models.SynapseNotebookReference :param parameters: Notebook parameters. :type parameters: dict[str, any] """ @@ -38691,7 +38957,7 @@ class SynapseNotebookReference(msrest.serialization.Model): :param type: Required. Synapse notebook reference type. Possible values include: "NotebookReference". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookReferenceType + :type type: str or ~azure.synapse.artifacts.models.NotebookReferenceType :param reference_name: Required. Reference notebook name. :type reference_name: str """ @@ -38733,16 +38999,15 @@ class SynapseSparkJobDefinitionActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param spark_job: Required. Synapse spark job reference. - :type spark_job: ~azure.synapse.artifacts.v2021_06_01_preview.models.SynapseSparkJobReference + :type spark_job: ~azure.synapse.artifacts.models.SynapseSparkJobReference :param arguments: User specified arguments to SynapseSparkJobDefinitionActivity. :type arguments: list[any] """ @@ -38793,7 +39058,7 @@ class SynapseSparkJobReference(msrest.serialization.Model): :param type: Required. Synapse spark job reference type. Possible values include: "SparkJobDefinitionReference". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobReferenceType + :type type: str or ~azure.synapse.artifacts.models.SparkJobReferenceType :param reference_name: Required. Reference spark job name. :type reference_name: str """ @@ -38855,8 +39120,7 @@ class TabularTranslator(CopyTranslator): activity. Type: boolean (or Expression with resultType boolean). :type type_conversion: any :param type_conversion_settings: Type conversion settings. - :type type_conversion_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.TypeConversionSettings + :type type_conversion_settings: ~azure.synapse.artifacts.models.TypeConversionSettings """ _validation = { @@ -38984,13 +39248,11 @@ class TeradataLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: Teradata ODBC connection string. Type: string, SecureString or @@ -39000,13 +39262,12 @@ class TeradataLinkedService(LinkedService): :type server: any :param authentication_type: AuthenticationType to be used for connection. Possible values include: "Basic", "Windows". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TeradataAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.TeradataAuthenticationType :param username: Username for authentication. Type: string (or Expression with resultType string). :type username: any :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -39123,12 +39384,10 @@ class TeradataSource(TabularSource): :type query: any :param partition_option: The partition mechanism that will be used for teradata read in parallel. Possible values include: "None", "Hash", "DynamicRange". - :type partition_option: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TeradataPartitionOption + :type partition_option: str or ~azure.synapse.artifacts.models.TeradataPartitionOption :param partition_settings: The settings that will be leveraged for teradata source partitioning. - :type partition_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.TeradataPartitionSettings + :type partition_settings: ~azure.synapse.artifacts.models.TeradataPartitionSettings """ _validation = { @@ -39188,16 +39447,14 @@ class TeradataTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param database: The database name of Teradata. Type: string (or Expression with resultType string). :type database: any @@ -39382,7 +39639,7 @@ class TriggerDependencyReference(DependencyReference): :param type: Required. The type of dependency reference.Constant filled by server. :type type: str :param reference_trigger: Required. Referenced trigger. - :type reference_trigger: ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerReference + :type reference_trigger: ~azure.synapse.artifacts.models.TriggerReference """ _validation = { @@ -39416,7 +39673,7 @@ class TriggerListResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of triggers. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerResource] + :type value: list[~azure.synapse.artifacts.models.TriggerResource] :param next_link: The link to the next page of results, if any remaining results exist. :type next_link: str """ @@ -39446,7 +39703,7 @@ class TriggerPipelineReference(msrest.serialization.Model): """Pipeline that needs to be triggered with the given parameters. :param pipeline_reference: Pipeline reference. - :type pipeline_reference: ~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineReference + :type pipeline_reference: ~azure.synapse.artifacts.models.PipelineReference :param parameters: Pipeline parameters. :type parameters: dict[str, any] """ @@ -39474,7 +39731,7 @@ class TriggerReference(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param type: Required. Trigger reference type. Possible values include: "TriggerReference". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerReferenceType + :type type: str or ~azure.synapse.artifacts.models.TriggerReferenceType :param reference_name: Required. Reference trigger name. :type reference_name: str """ @@ -39519,7 +39776,7 @@ class TriggerResource(SubResource): :ivar etag: Resource Etag. :vartype etag: str :param properties: Required. Properties of the trigger. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.Trigger + :type properties: ~azure.synapse.artifacts.models.Trigger """ _validation = { @@ -39565,7 +39822,7 @@ class TriggerRun(msrest.serialization.Model): :ivar trigger_run_timestamp: Trigger run start time. :vartype trigger_run_timestamp: ~datetime.datetime :ivar status: Trigger run status. Possible values include: "Succeeded", "Failed", "Inprogress". - :vartype status: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRunStatus + :vartype status: str or ~azure.synapse.artifacts.models.TriggerRunStatus :ivar message: Trigger error message. :vartype message: str :ivar properties: List of property name and value related to trigger run. Name, value pair @@ -39622,7 +39879,7 @@ class TriggerRunsQueryResponse(msrest.serialization.Model): All required parameters must be populated in order to send to Azure. :param value: Required. List of trigger runs. - :type value: list[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRun] + :type value: list[~azure.synapse.artifacts.models.TriggerRun] :param continuation_token: The continuation token for getting the next page of results, if any remaining results exist, null otherwise. :type continuation_token: str @@ -39658,8 +39915,7 @@ class TriggerSubscriptionOperationStatus(msrest.serialization.Model): :vartype trigger_name: str :ivar status: Event Subscription Status. Possible values include: "Enabled", "Provisioning", "Deprovisioning", "Disabled", "Unknown". - :vartype status: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.EventSubscriptionStatus + :vartype status: str or ~azure.synapse.artifacts.models.EventSubscriptionStatus """ _validation = { @@ -39697,17 +39953,15 @@ class TumblingWindowTrigger(Trigger): :type description: str :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRuntimeState + :vartype runtime_state: str or ~azure.synapse.artifacts.models.TriggerRuntimeState :param annotations: List of tags that can be used for describing the trigger. :type annotations: list[any] :param pipeline: Required. Pipeline for which runs are created when an event is fired for trigger window that is ready. - :type pipeline: ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerPipelineReference + :type pipeline: ~azure.synapse.artifacts.models.TriggerPipelineReference :param frequency: Required. The frequency of the time windows. Possible values include: "Minute", "Hour", "Month". - :type frequency: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.TumblingWindowFrequency + :type frequency: str or ~azure.synapse.artifacts.models.TumblingWindowFrequency :param interval: Required. The interval of the time windows. The minimum interval allowed is 15 Minutes. :type interval: int @@ -39725,10 +39979,10 @@ class TumblingWindowTrigger(Trigger): for which a new run is triggered. :type max_concurrency: int :param retry_policy: Retry policy that will be applied for failed pipeline runs. - :type retry_policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.RetryPolicy + :type retry_policy: ~azure.synapse.artifacts.models.RetryPolicy :param depends_on: Triggers that this trigger depends on. Only tumbling window triggers are supported. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DependencyReference] + :type depends_on: list[~azure.synapse.artifacts.models.DependencyReference] """ _validation = { @@ -39796,7 +40050,7 @@ class TumblingWindowTriggerDependencyReference(TriggerDependencyReference): :param type: Required. The type of dependency reference.Constant filled by server. :type type: str :param reference_trigger: Required. Referenced trigger. - :type reference_trigger: ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerReference + :type reference_trigger: ~azure.synapse.artifacts.models.TriggerReference :param offset: Timespan applied to the start time of a tumbling window when evaluating dependency. :type offset: str @@ -39900,12 +40154,12 @@ class UntilActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param expression: Required. An expression that would evaluate to Boolean. The loop will continue until this expression evaluates to true. - :type expression: ~azure.synapse.artifacts.v2021_06_01_preview.models.Expression + :type expression: ~azure.synapse.artifacts.models.Expression :param timeout: Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or Expression with resultType string), pattern: @@ -39913,7 +40167,7 @@ class UntilActivity(ControlActivity): resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type timeout: any :param activities: Required. List of activities to execute. - :type activities: list[~azure.synapse.artifacts.v2021_06_01_preview.models.Activity] + :type activities: list[~azure.synapse.artifacts.models.Activity] """ _validation = { @@ -40004,9 +40258,9 @@ class ValidationActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param timeout: Specifies the timeout for the activity to run. If there is no value specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or Expression with resultType string), pattern: @@ -40023,7 +40277,7 @@ class ValidationActivity(ControlActivity): with resultType boolean). :type child_items: any :param dataset: Required. Validation activity dataset reference. - :type dataset: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference + :type dataset: ~azure.synapse.artifacts.models.DatasetReference """ _validation = { @@ -40077,7 +40331,7 @@ class VariableSpecification(msrest.serialization.Model): :param type: Required. Variable type. Possible values include: "String", "Bool", "Boolean", "Array". - :type type: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.VariableType + :type type: str or ~azure.synapse.artifacts.models.VariableType :param default_value: Default value of variable. :type default_value: any """ @@ -40114,20 +40368,18 @@ class VerticaLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :type connection_string: any :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2021_06_01_preview.models.AzureKeyVaultSecretReference + :type pwd: ~azure.synapse.artifacts.models.AzureKeyVaultSecretReference :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string (or Expression with resultType string). @@ -40251,16 +40503,14 @@ class VerticaTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: This property will be retired. Please consider using schema + table properties instead. :type table_name: any @@ -40351,9 +40601,9 @@ class WaitActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param wait_time_in_seconds: Required. Duration in seconds. :type wait_time_in_seconds: any """ @@ -40405,17 +40655,16 @@ class WebActivity(ExecutionActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param linked_service_name: Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityPolicy + :type policy: ~azure.synapse.artifacts.models.ActivityPolicy :param method: Required. Rest API method for target endpoint. Possible values include: "GET", "POST", "PUT", "DELETE". - :type method: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.WebActivityMethod + :type method: str or ~azure.synapse.artifacts.models.WebActivityMethod :param url: Required. Web activity target endpoint and path. Type: string (or Expression with resultType string). :type url: any @@ -40427,16 +40676,13 @@ class WebActivity(ExecutionActivity): method, not allowed for GET method Type: string (or Expression with resultType string). :type body: any :param authentication: Authentication method used for calling the endpoint. - :type authentication: - ~azure.synapse.artifacts.v2021_06_01_preview.models.WebActivityAuthentication + :type authentication: ~azure.synapse.artifacts.models.WebActivityAuthentication :param datasets: List of datasets passed to web endpoint. - :type datasets: list[~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetReference] + :type datasets: list[~azure.synapse.artifacts.models.DatasetReference] :param linked_services: List of linked services passed to web endpoint. - :type linked_services: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference] + :type linked_services: list[~azure.synapse.artifacts.models.LinkedServiceReference] :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference """ _validation = { @@ -40505,11 +40751,11 @@ class WebActivityAuthentication(msrest.serialization.Model): :param type: Required. Web activity authentication (Basic/ClientCertificate/MSI). :type type: str :param pfx: Base64-encoded contents of a PFX file. - :type pfx: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type pfx: ~azure.synapse.artifacts.models.SecretBase :param username: Web activity authentication user name for basic authentication. :type username: str :param password: Password for the PFX file or basic authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase :param resource: Resource for which Azure Auth token will be requested when using MSI Authentication. :type resource: str @@ -40559,8 +40805,7 @@ class WebLinkedServiceTypeProperties(msrest.serialization.Model): :param authentication_type: Required. Type of authentication used to connect to the web table source.Constant filled by server. Possible values include: "Basic", "Anonymous", "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.WebAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.WebAuthenticationType """ _validation = { @@ -40599,8 +40844,7 @@ class WebAnonymousAuthentication(WebLinkedServiceTypeProperties): :param authentication_type: Required. Type of authentication used to connect to the web table source.Constant filled by server. Possible values include: "Basic", "Anonymous", "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.WebAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.WebAuthenticationType """ _validation = { @@ -40634,13 +40878,12 @@ class WebBasicAuthentication(WebLinkedServiceTypeProperties): :param authentication_type: Required. Type of authentication used to connect to the web table source.Constant filled by server. Possible values include: "Basic", "Anonymous", "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.WebAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.WebAuthenticationType :param username: Required. User name for Basic authentication. Type: string (or Expression with resultType string). :type username: any :param password: Required. The password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase """ _validation = { @@ -40682,12 +40925,11 @@ class WebClientCertificateAuthentication(WebLinkedServiceTypeProperties): :param authentication_type: Required. Type of authentication used to connect to the web table source.Constant filled by server. Possible values include: "Basic", "Anonymous", "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2021_06_01_preview.models.WebAuthenticationType + :type authentication_type: str or ~azure.synapse.artifacts.models.WebAuthenticationType :param pfx: Required. Base64-encoded contents of a PFX file. - :type pfx: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type pfx: ~azure.synapse.artifacts.models.SecretBase :param password: Required. Password for the PFX file. - :type password: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type password: ~azure.synapse.artifacts.models.SecretBase """ _validation = { @@ -40733,11 +40975,11 @@ class WebHookActivity(ControlActivity): :param description: Activity description. :type description: str :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityDependency] + :type depends_on: list[~azure.synapse.artifacts.models.ActivityDependency] :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2021_06_01_preview.models.UserProperty] + :type user_properties: list[~azure.synapse.artifacts.models.UserProperty] :param method: Required. Rest API method for target endpoint. Possible values include: "POST". - :type method: str or ~azure.synapse.artifacts.v2021_06_01_preview.models.WebHookActivityMethod + :type method: str or ~azure.synapse.artifacts.models.WebHookActivityMethod :param url: Required. WebHook activity target endpoint and path. Type: string (or Expression with resultType string). :type url: any @@ -40753,8 +40995,7 @@ class WebHookActivity(ControlActivity): method, not allowed for GET method Type: string (or Expression with resultType string). :type body: any :param authentication: Authentication method used for calling the endpoint. - :type authentication: - ~azure.synapse.artifacts.v2021_06_01_preview.models.WebActivityAuthentication + :type authentication: ~azure.synapse.artifacts.models.WebActivityAuthentication :param report_status_on_call_back: When set to true, statusCode, output and error in callback request body will be consumed by activity. The activity can be marked as failed by setting statusCode >= 400 in callback request. Default is false. Type: boolean (or Expression with @@ -40824,18 +41065,15 @@ class WebLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param type_properties: Required. Web linked service properties. - :type type_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.WebLinkedServiceTypeProperties + :type type_properties: ~azure.synapse.artifacts.models.WebLinkedServiceTypeProperties """ _validation = { @@ -40940,16 +41178,14 @@ class WebTableDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param index: Required. The zero-based index of the table in the web page. Type: integer (or Expression with resultType integer), minimum: 0. :type index: any @@ -41019,10 +41255,9 @@ class Workspace(TrackedResource): :param location: Required. The geo-location where the resource lives. :type location: str :param identity: Identity of the workspace. - :type identity: ~azure.synapse.artifacts.v2021_06_01_preview.models.ManagedIdentity + :type identity: ~azure.synapse.artifacts.models.ManagedIdentity :param default_data_lake_storage: Workspace default data lake storage account details. - :type default_data_lake_storage: - ~azure.synapse.artifacts.v2021_06_01_preview.models.DataLakeStorageAccountDetails + :type default_data_lake_storage: ~azure.synapse.artifacts.models.DataLakeStorageAccountDetails :param sql_administrator_login_password: SQL administrator login password. :type sql_administrator_login_password: str :param managed_resource_group_name: Workspace managed resource group. The resource group name @@ -41035,8 +41270,7 @@ class Workspace(TrackedResource): :param sql_administrator_login: Login for workspace SQL active directory administrator. :type sql_administrator_login: str :param virtual_network_profile: Virtual Network profile. - :type virtual_network_profile: - ~azure.synapse.artifacts.v2021_06_01_preview.models.VirtualNetworkProfile + :type virtual_network_profile: ~azure.synapse.artifacts.models.VirtualNetworkProfile :param connectivity_endpoints: Connectivity endpoints. :type connectivity_endpoints: dict[str, str] :param managed_virtual_network: Setting this to 'default' will ensure that all compute for this @@ -41044,22 +41278,21 @@ class Workspace(TrackedResource): :type managed_virtual_network: str :param private_endpoint_connections: Private endpoint connections to the workspace. :type private_endpoint_connections: - list[~azure.synapse.artifacts.v2021_06_01_preview.models.PrivateEndpointConnection] + list[~azure.synapse.artifacts.models.PrivateEndpointConnection] :param encryption: The encryption details of the workspace. - :type encryption: ~azure.synapse.artifacts.v2021_06_01_preview.models.EncryptionDetails + :type encryption: ~azure.synapse.artifacts.models.EncryptionDetails :ivar workspace_uid: The workspace unique identifier. :vartype workspace_uid: str :ivar extra_properties: Workspace level configs and feature flags. :vartype extra_properties: dict[str, any] :param managed_virtual_network_settings: Managed Virtual Network Settings. :type managed_virtual_network_settings: - ~azure.synapse.artifacts.v2021_06_01_preview.models.ManagedVirtualNetworkSettings + ~azure.synapse.artifacts.models.ManagedVirtualNetworkSettings :param workspace_repository_configuration: Git integration settings. :type workspace_repository_configuration: - ~azure.synapse.artifacts.v2021_06_01_preview.models.WorkspaceRepositoryConfiguration + ~azure.synapse.artifacts.models.WorkspaceRepositoryConfiguration :param purview_configuration: Purview Configuration. - :type purview_configuration: - ~azure.synapse.artifacts.v2021_06_01_preview.models.PurviewConfiguration + :type purview_configuration: ~azure.synapse.artifacts.models.PurviewConfiguration :ivar adla_resource_id: The ADLA resource ID. :vartype adla_resource_id: str """ @@ -41230,7 +41463,7 @@ class WorkspaceRepositoryConfiguration(msrest.serialization.Model): :param client_id: GitHub bring your own app client id. :type client_id: str :param client_secret: GitHub bring your own app client secret information. - :type client_secret: ~azure.synapse.artifacts.v2021_06_01_preview.models.GitHubClientSecret + :type client_secret: ~azure.synapse.artifacts.models.GitHubClientSecret """ _attribute_map = { @@ -41283,7 +41516,7 @@ class WorkspaceUpdateParameters(msrest.serialization.Model): :param tags: A set of tags. The resource tags. :type tags: dict[str, str] :param identity: Managed service identity of the workspace. - :type identity: ~azure.synapse.artifacts.v2021_06_01_preview.models.WorkspaceIdentity + :type identity: ~azure.synapse.artifacts.models.WorkspaceIdentity """ _attribute_map = { @@ -41314,13 +41547,11 @@ class XeroLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_properties: Properties used to connect to Xero. It is mutually exclusive with @@ -41329,11 +41560,11 @@ class XeroLinkedService(LinkedService): :param host: Required. The endpoint of the Xero server. (i.e. api.xero.com). :type host: any :param consumer_key: The consumer key associated with the Xero application. - :type consumer_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type consumer_key: ~azure.synapse.artifacts.models.SecretBase :param private_key: The private key from the .pem file that was generated for your Xero private application. You must include all the text from the .pem file, including the Unix line endings( ). - :type private_key: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type private_key: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -41421,16 +41652,14 @@ class XeroObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ @@ -41553,18 +41782,16 @@ class XmlDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param location: The location of the json data storage. - :type location: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetLocation + :type location: ~azure.synapse.artifacts.models.DatasetLocation :param encoding_name: The code page name of the preferred encoding. If not specified, the default value is UTF-8, unless BOM denotes another Unicode encoding. Refer to the name column of the table in the following link to set supported values: @@ -41574,7 +41801,7 @@ class XmlDataset(Dataset): :param null_value: The null value string. Type: string (or Expression with resultType string). :type null_value: any :param compression: The data compression method used for the json dataset. - :type compression: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetCompression + :type compression: ~azure.synapse.artifacts.models.DatasetCompression """ _validation = { @@ -41634,8 +41861,7 @@ class XmlReadSettings(FormatReadSettings): :param type: Required. The read setting type.Constant filled by server. :type type: str :param compression_properties: Compression settings. - :type compression_properties: - ~azure.synapse.artifacts.v2021_06_01_preview.models.CompressionReadSettings + :type compression_properties: ~azure.synapse.artifacts.models.CompressionReadSettings :param validation_mode: Indicates what validation method is used when reading the xml files. Allowed values: 'none', 'xsd', or 'dtd'. Type: string (or Expression with resultType string). :type validation_mode: any @@ -41706,9 +41932,9 @@ class XmlSource(CopySource): store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: any :param store_settings: Xml store settings. - :type store_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.StoreReadSettings + :type store_settings: ~azure.synapse.artifacts.models.StoreReadSettings :param format_settings: Xml format settings. - :type format_settings: ~azure.synapse.artifacts.v2021_06_01_preview.models.XmlReadSettings + :type format_settings: ~azure.synapse.artifacts.models.XmlReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :type additional_columns: any @@ -41796,13 +42022,11 @@ class ZohoLinkedService(LinkedService): :param type: Required. Type of linked service.Constant filled by server. :type type: str :param connect_via: The integration runtime reference. - :type connect_via: - ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeReference + :type connect_via: ~azure.synapse.artifacts.models.IntegrationRuntimeReference :param description: Linked service description. :type description: str :param parameters: Parameters for linked service. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] :param connection_properties: Properties used to connect to Zoho. It is mutually exclusive with @@ -41811,7 +42035,7 @@ class ZohoLinkedService(LinkedService): :param endpoint: Required. The endpoint of the Zoho server. (i.e. crm.zoho.com/crm/private). :type endpoint: any :param access_token: The access token for Zoho authentication. - :type access_token: ~azure.synapse.artifacts.v2021_06_01_preview.models.SecretBase + :type access_token: ~azure.synapse.artifacts.models.SecretBase :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using HTTPS. The default value is true. :type use_encrypted_endpoints: any @@ -41896,16 +42120,14 @@ class ZohoObjectDataset(Dataset): Expression with resultType array), itemType: DatasetSchemaDataElement. :type schema: any :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceReference + :type linked_service_name: ~azure.synapse.artifacts.models.LinkedServiceReference :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.synapse.artifacts.v2021_06_01_preview.models.ParameterSpecification] + :type parameters: dict[str, ~azure.synapse.artifacts.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the Dataset. :type annotations: list[any] :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - :type folder: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetFolder + :type folder: ~azure.synapse.artifacts.models.DatasetFolder :param table_name: The table name. Type: string (or Expression with resultType string). :type table_name: any """ diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/__init__.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/__init__.py similarity index 91% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/__init__.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/__init__.py index 4620c478f14b..7b2bd054134a 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/__init__.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/__init__.py @@ -6,48 +6,48 @@ # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -from ._linked_service_operations import LinkedServiceOperations -from ._dataset_operations import DatasetOperations -from ._pipeline_operations import PipelineOperations -from ._pipeline_run_operations import PipelineRunOperations -from ._trigger_operations import TriggerOperations -from ._trigger_run_operations import TriggerRunOperations -from ._data_flow_operations import DataFlowOperations -from ._data_flow_debug_session_operations import DataFlowDebugSessionOperations -from ._sql_script_operations import SqlScriptOperations -from ._spark_job_definition_operations import SparkJobDefinitionOperations -from ._notebook_operations import NotebookOperations -from ._notebook_operation_result_operations import NotebookOperationResultOperations +from ._kql_scripts_operations import KqlScriptsOperations +from ._kql_script_operations import KqlScriptOperations from ._spark_configuration_operations import SparkConfigurationOperations from ._big_data_pools_operations import BigDataPoolsOperations +from ._data_flow_operations import DataFlowOperations +from ._data_flow_debug_session_operations import DataFlowDebugSessionOperations +from ._dataset_operations import DatasetOperations from ._workspace_git_repo_management_operations import WorkspaceGitRepoManagementOperations from ._integration_runtimes_operations import IntegrationRuntimesOperations from ._library_operations import LibraryOperations -from ._operation_result_operations import OperationResultOperations -from ._operation_status_operations import OperationStatusOperations +from ._linked_service_operations import LinkedServiceOperations +from ._notebook_operations import NotebookOperations +from ._notebook_operation_result_operations import NotebookOperationResultOperations +from ._pipeline_operations import PipelineOperations +from ._pipeline_run_operations import PipelineRunOperations +from ._spark_job_definition_operations import SparkJobDefinitionOperations from ._sql_pools_operations import SqlPoolsOperations +from ._sql_script_operations import SqlScriptOperations +from ._trigger_operations import TriggerOperations +from ._trigger_run_operations import TriggerRunOperations from ._workspace_operations import WorkspaceOperations __all__ = [ - 'LinkedServiceOperations', - 'DatasetOperations', - 'PipelineOperations', - 'PipelineRunOperations', - 'TriggerOperations', - 'TriggerRunOperations', - 'DataFlowOperations', - 'DataFlowDebugSessionOperations', - 'SqlScriptOperations', - 'SparkJobDefinitionOperations', - 'NotebookOperations', - 'NotebookOperationResultOperations', + 'KqlScriptsOperations', + 'KqlScriptOperations', 'SparkConfigurationOperations', 'BigDataPoolsOperations', + 'DataFlowOperations', + 'DataFlowDebugSessionOperations', + 'DatasetOperations', 'WorkspaceGitRepoManagementOperations', 'IntegrationRuntimesOperations', 'LibraryOperations', - 'OperationResultOperations', - 'OperationStatusOperations', + 'LinkedServiceOperations', + 'NotebookOperations', + 'NotebookOperationResultOperations', + 'PipelineOperations', + 'PipelineRunOperations', + 'SparkJobDefinitionOperations', 'SqlPoolsOperations', + 'SqlScriptOperations', + 'TriggerOperations', + 'TriggerRunOperations', 'WorkspaceOperations', ] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_big_data_pools_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_big_data_pools_operations.py similarity index 96% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_big_data_pools_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_big_data_pools_operations.py index a7ee6bc8aaf3..da48714f818e 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_big_data_pools_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_big_data_pools_operations.py @@ -28,7 +28,7 @@ class BigDataPoolsOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -52,7 +52,7 @@ def list( :keyword callable cls: A custom type or function that will be passed the direct response :return: BigDataPoolResourceInfoListResult, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.BigDataPoolResourceInfoListResult + :rtype: ~azure.synapse.artifacts.models.BigDataPoolResourceInfoListResult :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.BigDataPoolResourceInfoListResult"] @@ -107,7 +107,7 @@ def get( :type big_data_pool_name: str :keyword callable cls: A custom type or function that will be passed the direct response :return: BigDataPoolResourceInfo, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.BigDataPoolResourceInfo + :rtype: ~azure.synapse.artifacts.models.BigDataPoolResourceInfo :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.BigDataPoolResourceInfo"] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_data_flow_debug_session_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_data_flow_debug_session_operations.py similarity index 96% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_data_flow_debug_session_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_data_flow_debug_session_operations.py index 2e1801cf9897..5df441805270 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_data_flow_debug_session_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_data_flow_debug_session_operations.py @@ -31,7 +31,7 @@ class DataFlowDebugSessionOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -86,7 +86,7 @@ def _create_data_flow_debug_session_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) response_headers = {} @@ -112,7 +112,7 @@ def begin_create_data_flow_debug_session( """Creates a data flow debug session. :param request: Data flow debug session definition. - :type request: ~azure.synapse.artifacts.v2020_12_01.models.CreateDataFlowDebugSessionRequest + :type request: ~azure.synapse.artifacts.models.CreateDataFlowDebugSessionRequest :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: By default, your polling method will be LROBasePolling. @@ -120,7 +120,7 @@ def begin_create_data_flow_debug_session( :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either CreateDataFlowDebugSessionResponse or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2020_12_01.models.CreateDataFlowDebugSessionResponse] + :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.models.CreateDataFlowDebugSessionResponse] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] @@ -174,7 +174,7 @@ def query_data_flow_debug_sessions_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either QueryDataFlowDebugSessionsResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2020_12_01.models.QueryDataFlowDebugSessionsResponse] + :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.models.QueryDataFlowDebugSessionsResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.QueryDataFlowDebugSessionsResponse"] @@ -226,7 +226,7 @@ def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -246,10 +246,10 @@ def add_data_flow( """Add a data flow into debug session. :param request: Data flow debug session definition with debug content. - :type request: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowDebugPackage + :type request: ~azure.synapse.artifacts.models.DataFlowDebugPackage :keyword callable cls: A custom type or function that will be passed the direct response :return: AddDataFlowToDebugSessionResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.AddDataFlowToDebugSessionResponse + :rtype: ~azure.synapse.artifacts.models.AddDataFlowToDebugSessionResponse :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.AddDataFlowToDebugSessionResponse"] @@ -286,7 +286,7 @@ def add_data_flow( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = self._deserialize('AddDataFlowToDebugSessionResponse', pipeline_response) @@ -306,7 +306,7 @@ def delete_data_flow_debug_session( """Deletes a data flow debug session. :param request: Data flow debug session definition for deletion. - :type request: ~azure.synapse.artifacts.v2020_12_01.models.DeleteDataFlowDebugSessionRequest + :type request: ~azure.synapse.artifacts.models.DeleteDataFlowDebugSessionRequest :keyword callable cls: A custom type or function that will be passed the direct response :return: None, or the result of cls(response) :rtype: None @@ -346,7 +346,7 @@ def delete_data_flow_debug_session( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -394,7 +394,7 @@ def _execute_command_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) response_headers = {} @@ -420,7 +420,7 @@ def begin_execute_command( """Execute a data flow debug command. :param request: Data flow debug command definition. - :type request: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowDebugCommandRequest + :type request: ~azure.synapse.artifacts.models.DataFlowDebugCommandRequest :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: By default, your polling method will be LROBasePolling. @@ -428,7 +428,7 @@ def begin_execute_command( :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either DataFlowDebugCommandResponse or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2020_12_01.models.DataFlowDebugCommandResponse] + :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.models.DataFlowDebugCommandResponse] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_data_flow_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_data_flow_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_data_flow_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_data_flow_operations.py index 0409cef82120..cdf9f6e756a7 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_data_flow_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_data_flow_operations.py @@ -31,7 +31,7 @@ class DataFlowOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -93,7 +93,7 @@ def _create_or_update_data_flow_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -119,7 +119,7 @@ def begin_create_or_update_data_flow( :param data_flow_name: The data flow name. :type data_flow_name: str :param properties: Data flow properties. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.DataFlow + :type properties: ~azure.synapse.artifacts.models.DataFlow :param if_match: ETag of the data flow entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -130,7 +130,7 @@ def begin_create_or_update_data_flow( :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either DataFlowResource or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2020_12_01.models.DataFlowResource] + :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.models.DataFlowResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] @@ -194,7 +194,7 @@ def get_data_flow( :type if_none_match: str :keyword callable cls: A custom type or function that will be passed the direct response :return: DataFlowResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowResource + :rtype: ~azure.synapse.artifacts.models.DataFlowResource :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.DataFlowResource"] @@ -229,7 +229,7 @@ def get_data_flow( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = self._deserialize('DataFlowResource', pipeline_response) @@ -276,7 +276,7 @@ def _delete_data_flow_initial( if response.status_code not in [200, 202, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -388,7 +388,7 @@ def _rename_data_flow_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -469,7 +469,7 @@ def get_data_flows_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either DataFlowListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2020_12_01.models.DataFlowListResponse] + :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.models.DataFlowListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.DataFlowListResponse"] @@ -521,7 +521,7 @@ def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_dataset_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_dataset_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_dataset_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_dataset_operations.py index 4289d3c944d3..d060ddee098d 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_dataset_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_dataset_operations.py @@ -31,7 +31,7 @@ class DatasetOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -55,7 +55,7 @@ def get_datasets_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either DatasetListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2020_12_01.models.DatasetListResponse] + :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.models.DatasetListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.DatasetListResponse"] @@ -107,7 +107,7 @@ def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -165,7 +165,7 @@ def _create_or_update_dataset_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -191,7 +191,7 @@ def begin_create_or_update_dataset( :param dataset_name: The dataset name. :type dataset_name: str :param properties: Dataset properties. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.Dataset + :type properties: ~azure.synapse.artifacts.models.Dataset :param if_match: ETag of the dataset entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -202,7 +202,7 @@ def begin_create_or_update_dataset( :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either DatasetResource or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2020_12_01.models.DatasetResource] + :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.models.DatasetResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] @@ -266,7 +266,7 @@ def get_dataset( :type if_none_match: str :keyword callable cls: A custom type or function that will be passed the direct response :return: DatasetResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.DatasetResource or None + :rtype: ~azure.synapse.artifacts.models.DatasetResource or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.DatasetResource"]] @@ -301,7 +301,7 @@ def get_dataset( if response.status_code not in [200, 304]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -350,7 +350,7 @@ def _delete_dataset_initial( if response.status_code not in [200, 202, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -462,7 +462,7 @@ def _rename_dataset_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_integration_runtimes_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_integration_runtimes_operations.py similarity index 96% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_integration_runtimes_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_integration_runtimes_operations.py index 86e5a764988d..4a7729772eb3 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_integration_runtimes_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_integration_runtimes_operations.py @@ -28,7 +28,7 @@ class IntegrationRuntimesOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -52,7 +52,7 @@ def list( :keyword callable cls: A custom type or function that will be passed the direct response :return: IntegrationRuntimeListResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeListResponse + :rtype: ~azure.synapse.artifacts.models.IntegrationRuntimeListResponse :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.IntegrationRuntimeListResponse"] @@ -107,7 +107,7 @@ def get( :type integration_runtime_name: str :keyword callable cls: A custom type or function that will be passed the direct response :return: IntegrationRuntimeResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeResource + :rtype: ~azure.synapse.artifacts.models.IntegrationRuntimeResource :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.IntegrationRuntimeResource"] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_kql_script_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_kql_script_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_kql_script_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_kql_script_operations.py index e8efde19cf63..39858d206044 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_kql_script_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_kql_script_operations.py @@ -30,7 +30,7 @@ class KqlScriptOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -112,7 +112,7 @@ def begin_create_or_update( :param kql_script_name: KQL script name. :type kql_script_name: str :param kql_script: KQL script. - :type kql_script: ~azure.synapse.artifacts.v2021_06_01_preview.models.KqlScriptResource + :type kql_script: ~azure.synapse.artifacts.models.KqlScriptResource :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: By default, your polling method will be LROBasePolling. @@ -120,7 +120,7 @@ def begin_create_or_update( :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either KqlScriptResource or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.KqlScriptResource] + :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.models.KqlScriptResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] @@ -179,7 +179,7 @@ def get_by_name( :type kql_script_name: str :keyword callable cls: A custom type or function that will be passed the direct response :return: KqlScriptResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.KqlScriptResource + :rtype: ~azure.synapse.artifacts.models.KqlScriptResource :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.KqlScriptResource"] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_kql_scripts_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_kql_scripts_operations.py similarity index 97% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_kql_scripts_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_kql_scripts_operations.py index 01f3426cc8fa..71bd2113da2a 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_kql_scripts_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_kql_scripts_operations.py @@ -29,7 +29,7 @@ class KqlScriptsOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -53,7 +53,7 @@ def get_all( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either KqlScriptsResourceCollectionResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.KqlScriptsResourceCollectionResponse] + :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.models.KqlScriptsResourceCollectionResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.KqlScriptsResourceCollectionResponse"] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_library_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_library_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_library_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_library_operations.py index c587c77105d1..c6686dcd2312 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_library_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_library_operations.py @@ -31,7 +31,7 @@ class LibraryOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -55,7 +55,7 @@ def list( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either LibraryListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2020_12_01.models.LibraryListResponse] + :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.models.LibraryListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.LibraryListResponse"] @@ -107,7 +107,7 @@ def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -154,7 +154,7 @@ def _flush_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -243,7 +243,7 @@ def get_operation_result( :type operation_id: str :keyword callable cls: A custom type or function that will be passed the direct response :return: LibraryResource or OperationResult, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.LibraryResource or ~azure.synapse.artifacts.v2020_12_01.models.OperationResult + :rtype: ~azure.synapse.artifacts.models.LibraryResource or ~azure.synapse.artifacts.models.OperationResult :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Union["_models.LibraryResource", "_models.OperationResult"]] @@ -276,7 +276,7 @@ def get_operation_result( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if response.status_code == 200: @@ -327,7 +327,7 @@ def _delete_initial( if response.status_code not in [200, 202, 409]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -417,7 +417,7 @@ def get( :type library_name: str :keyword callable cls: A custom type or function that will be passed the direct response :return: LibraryResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.LibraryResource or None + :rtype: ~azure.synapse.artifacts.models.LibraryResource or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.LibraryResource"]] @@ -450,7 +450,7 @@ def get( if response.status_code not in [200, 304]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -499,7 +499,7 @@ def _create_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -640,7 +640,7 @@ def append( if response.status_code not in [201]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_linked_service_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_linked_service_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_linked_service_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_linked_service_operations.py index c85af59bb331..d6c73e6a3bfc 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_linked_service_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_linked_service_operations.py @@ -31,7 +31,7 @@ class LinkedServiceOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -55,7 +55,7 @@ def get_linked_services_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either LinkedServiceListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceListResponse] + :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.models.LinkedServiceListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.LinkedServiceListResponse"] @@ -107,7 +107,7 @@ def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -165,7 +165,7 @@ def _create_or_update_linked_service_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -191,7 +191,7 @@ def begin_create_or_update_linked_service( :param linked_service_name: The linked service name. :type linked_service_name: str :param properties: Properties of linked service. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.LinkedService + :type properties: ~azure.synapse.artifacts.models.LinkedService :param if_match: ETag of the linkedService entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -202,7 +202,7 @@ def begin_create_or_update_linked_service( :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either LinkedServiceResource or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceResource] + :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.models.LinkedServiceResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] @@ -267,7 +267,7 @@ def get_linked_service( :type if_none_match: str :keyword callable cls: A custom type or function that will be passed the direct response :return: LinkedServiceResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceResource or None + :rtype: ~azure.synapse.artifacts.models.LinkedServiceResource or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.LinkedServiceResource"]] @@ -302,7 +302,7 @@ def get_linked_service( if response.status_code not in [200, 304]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -351,7 +351,7 @@ def _delete_linked_service_initial( if response.status_code not in [200, 202, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -463,7 +463,7 @@ def _rename_linked_service_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_notebook_operation_result_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_notebook_operation_result_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_notebook_operation_result_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_notebook_operation_result_operations.py index 2abb8a8b738c..55f6c1a6f1c8 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_notebook_operation_result_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_notebook_operation_result_operations.py @@ -28,7 +28,7 @@ class NotebookOperationResultOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_notebook_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_notebook_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_notebook_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_notebook_operations.py index 6b7a9e4fc261..51877617e146 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_notebook_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_notebook_operations.py @@ -31,7 +31,7 @@ class NotebookOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -55,7 +55,7 @@ def get_notebooks_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either NotebookListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2020_12_01.models.NotebookListResponse] + :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.models.NotebookListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.NotebookListResponse"] @@ -107,7 +107,7 @@ def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -127,7 +127,7 @@ def get_notebook_summary_by_work_space( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either NotebookListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2020_12_01.models.NotebookListResponse] + :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.models.NotebookListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.NotebookListResponse"] @@ -179,7 +179,7 @@ def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -235,7 +235,7 @@ def _create_or_update_notebook_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -261,7 +261,7 @@ def begin_create_or_update_notebook( :param notebook_name: The notebook name. :type notebook_name: str :param notebook: Note book resource definition. - :type notebook: ~azure.synapse.artifacts.v2020_12_01.models.NotebookResource + :type notebook: ~azure.synapse.artifacts.models.NotebookResource :param if_match: ETag of the Note book entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -272,7 +272,7 @@ def begin_create_or_update_notebook( :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either NotebookResource or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2020_12_01.models.NotebookResource] + :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.models.NotebookResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] @@ -336,7 +336,7 @@ def get_notebook( :type if_none_match: str :keyword callable cls: A custom type or function that will be passed the direct response :return: NotebookResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.NotebookResource or None + :rtype: ~azure.synapse.artifacts.models.NotebookResource or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.NotebookResource"]] @@ -371,7 +371,7 @@ def get_notebook( if response.status_code not in [200, 304]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -420,7 +420,7 @@ def _delete_notebook_initial( if response.status_code not in [200, 202, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -532,7 +532,7 @@ def _rename_notebook_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_pipeline_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_pipeline_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_pipeline_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_pipeline_operations.py index 68fdf39615b7..ae179cf17ce6 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_pipeline_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_pipeline_operations.py @@ -31,7 +31,7 @@ class PipelineOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -55,7 +55,7 @@ def get_pipelines_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either PipelineListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2020_12_01.models.PipelineListResponse] + :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.models.PipelineListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.PipelineListResponse"] @@ -107,7 +107,7 @@ def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -163,7 +163,7 @@ def _create_or_update_pipeline_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -189,7 +189,7 @@ def begin_create_or_update_pipeline( :param pipeline_name: The pipeline name. :type pipeline_name: str :param pipeline: Pipeline resource definition. - :type pipeline: ~azure.synapse.artifacts.v2020_12_01.models.PipelineResource + :type pipeline: ~azure.synapse.artifacts.models.PipelineResource :param if_match: ETag of the pipeline entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -200,7 +200,7 @@ def begin_create_or_update_pipeline( :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either PipelineResource or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2020_12_01.models.PipelineResource] + :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.models.PipelineResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] @@ -264,7 +264,7 @@ def get_pipeline( :type if_none_match: str :keyword callable cls: A custom type or function that will be passed the direct response :return: PipelineResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.PipelineResource or None + :rtype: ~azure.synapse.artifacts.models.PipelineResource or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.PipelineResource"]] @@ -299,7 +299,7 @@ def get_pipeline( if response.status_code not in [200, 304]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -348,7 +348,7 @@ def _delete_pipeline_initial( if response.status_code not in [200, 202, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -460,7 +460,7 @@ def _rename_pipeline_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -560,7 +560,7 @@ def create_pipeline_run( :type parameters: dict[str, any] :keyword callable cls: A custom type or function that will be passed the direct response :return: CreateRunResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.CreateRunResponse + :rtype: ~azure.synapse.artifacts.models.CreateRunResponse :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.CreateRunResponse"] @@ -607,7 +607,7 @@ def create_pipeline_run( if response.status_code not in [202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = self._deserialize('CreateRunResponse', pipeline_response) diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_pipeline_run_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_pipeline_run_operations.py similarity index 95% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_pipeline_run_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_pipeline_run_operations.py index 77322f1d5991..2dc41f73c362 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_pipeline_run_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_pipeline_run_operations.py @@ -28,7 +28,7 @@ class PipelineRunOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -52,10 +52,10 @@ def query_pipeline_runs_by_workspace( """Query pipeline runs in the workspace based on input filter conditions. :param filter_parameters: Parameters to filter the pipeline run. - :type filter_parameters: ~azure.synapse.artifacts.v2020_12_01.models.RunFilterParameters + :type filter_parameters: ~azure.synapse.artifacts.models.RunFilterParameters :keyword callable cls: A custom type or function that will be passed the direct response :return: PipelineRunsQueryResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.PipelineRunsQueryResponse + :rtype: ~azure.synapse.artifacts.models.PipelineRunsQueryResponse :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.PipelineRunsQueryResponse"] @@ -92,7 +92,7 @@ def query_pipeline_runs_by_workspace( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = self._deserialize('PipelineRunsQueryResponse', pipeline_response) @@ -115,7 +115,7 @@ def get_pipeline_run( :type run_id: str :keyword callable cls: A custom type or function that will be passed the direct response :return: PipelineRun, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.PipelineRun + :rtype: ~azure.synapse.artifacts.models.PipelineRun :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.PipelineRun"] @@ -148,7 +148,7 @@ def get_pipeline_run( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = self._deserialize('PipelineRun', pipeline_response) @@ -174,10 +174,10 @@ def query_activity_runs( :param run_id: The pipeline run identifier. :type run_id: str :param filter_parameters: Parameters to filter the activity runs. - :type filter_parameters: ~azure.synapse.artifacts.v2020_12_01.models.RunFilterParameters + :type filter_parameters: ~azure.synapse.artifacts.models.RunFilterParameters :keyword callable cls: A custom type or function that will be passed the direct response :return: ActivityRunsQueryResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.ActivityRunsQueryResponse + :rtype: ~azure.synapse.artifacts.models.ActivityRunsQueryResponse :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.ActivityRunsQueryResponse"] @@ -216,7 +216,7 @@ def query_activity_runs( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = self._deserialize('ActivityRunsQueryResponse', pipeline_response) @@ -278,7 +278,7 @@ def cancel_pipeline_run( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_spark_configuration_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_spark_configuration_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_spark_configuration_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_spark_configuration_operations.py index 41c6706ba041..d47f8fd211a3 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_spark_configuration_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_spark_configuration_operations.py @@ -31,7 +31,7 @@ class SparkConfigurationOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -55,7 +55,7 @@ def get_spark_configurations_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either SparkConfigurationListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.SparkConfigurationListResponse] + :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.models.SparkConfigurationListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.SparkConfigurationListResponse"] @@ -191,7 +191,7 @@ def begin_create_or_update_spark_configuration( :param spark_configuration_name: The spark Configuration name. :type spark_configuration_name: str :param properties: Properties of Spark Configuration. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkConfiguration + :type properties: ~azure.synapse.artifacts.models.SparkConfiguration :param if_match: ETag of the sparkConfiguration entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -202,7 +202,7 @@ def begin_create_or_update_spark_configuration( :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either SparkConfigurationResource or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.SparkConfigurationResource] + :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.models.SparkConfigurationResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] @@ -267,7 +267,7 @@ def get_spark_configuration( :type if_none_match: str :keyword callable cls: A custom type or function that will be passed the direct response :return: SparkConfigurationResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkConfigurationResource or None + :rtype: ~azure.synapse.artifacts.models.SparkConfigurationResource or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.SparkConfigurationResource"]] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_spark_job_definition_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_spark_job_definition_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_spark_job_definition_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_spark_job_definition_operations.py index b73a6bfcfb48..9ebd9f782080 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_spark_job_definition_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_spark_job_definition_operations.py @@ -31,7 +31,7 @@ class SparkJobDefinitionOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -55,7 +55,7 @@ def get_spark_job_definitions_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either SparkJobDefinitionsListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2020_12_01.models.SparkJobDefinitionsListResponse] + :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.models.SparkJobDefinitionsListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.SparkJobDefinitionsListResponse"] @@ -107,7 +107,7 @@ def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -165,7 +165,7 @@ def _create_or_update_spark_job_definition_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -191,7 +191,7 @@ def begin_create_or_update_spark_job_definition( :param spark_job_definition_name: The spark job definition name. :type spark_job_definition_name: str :param properties: Properties of spark job definition. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.SparkJobDefinition + :type properties: ~azure.synapse.artifacts.models.SparkJobDefinition :param if_match: ETag of the Spark Job Definition entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -202,7 +202,7 @@ def begin_create_or_update_spark_job_definition( :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either SparkJobDefinitionResource or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2020_12_01.models.SparkJobDefinitionResource] + :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.models.SparkJobDefinitionResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] @@ -267,7 +267,7 @@ def get_spark_job_definition( :type if_none_match: str :keyword callable cls: A custom type or function that will be passed the direct response :return: SparkJobDefinitionResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.SparkJobDefinitionResource or None + :rtype: ~azure.synapse.artifacts.models.SparkJobDefinitionResource or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.SparkJobDefinitionResource"]] @@ -302,7 +302,7 @@ def get_spark_job_definition( if response.status_code not in [200, 304]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -351,7 +351,7 @@ def _delete_spark_job_definition_initial( if response.status_code not in [200, 202, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -455,7 +455,7 @@ def _execute_spark_job_definition_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if response.status_code == 200: @@ -487,7 +487,7 @@ def begin_execute_spark_job_definition( :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either SparkBatchJob or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2020_12_01.models.SparkBatchJob] + :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.models.SparkBatchJob] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] @@ -577,7 +577,7 @@ def _rename_spark_job_definition_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -691,7 +691,7 @@ def _debug_spark_job_definition_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if response.status_code == 200: @@ -715,7 +715,7 @@ def begin_debug_spark_job_definition( """Debug the spark job definition. :param properties: Properties of spark job definition. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.SparkJobDefinition + :type properties: ~azure.synapse.artifacts.models.SparkJobDefinition :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: By default, your polling method will be LROBasePolling. @@ -723,7 +723,7 @@ def begin_debug_spark_job_definition( :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either SparkBatchJob or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2020_12_01.models.SparkBatchJob] + :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.models.SparkBatchJob] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_sql_pools_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_sql_pools_operations.py similarity index 96% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_sql_pools_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_sql_pools_operations.py index 34a53dcbea50..ff85b93fb34f 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_sql_pools_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_sql_pools_operations.py @@ -28,7 +28,7 @@ class SqlPoolsOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -52,7 +52,7 @@ def list( :keyword callable cls: A custom type or function that will be passed the direct response :return: SqlPoolInfoListResult, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.SqlPoolInfoListResult + :rtype: ~azure.synapse.artifacts.models.SqlPoolInfoListResult :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.SqlPoolInfoListResult"] @@ -107,7 +107,7 @@ def get( :type sql_pool_name: str :keyword callable cls: A custom type or function that will be passed the direct response :return: SqlPool, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.SqlPool + :rtype: ~azure.synapse.artifacts.models.SqlPool :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.SqlPool"] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_sql_script_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_sql_script_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_sql_script_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_sql_script_operations.py index 4801813c58d4..7605c1fee2f8 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_sql_script_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_sql_script_operations.py @@ -31,7 +31,7 @@ class SqlScriptOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -55,7 +55,7 @@ def get_sql_scripts_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either SqlScriptsListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2020_12_01.models.SqlScriptsListResponse] + :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.models.SqlScriptsListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.SqlScriptsListResponse"] @@ -107,7 +107,7 @@ def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -163,7 +163,7 @@ def _create_or_update_sql_script_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -189,7 +189,7 @@ def begin_create_or_update_sql_script( :param sql_script_name: The sql script name. :type sql_script_name: str :param sql_script: Sql Script resource definition. - :type sql_script: ~azure.synapse.artifacts.v2020_12_01.models.SqlScriptResource + :type sql_script: ~azure.synapse.artifacts.models.SqlScriptResource :param if_match: ETag of the SQL script entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -200,7 +200,7 @@ def begin_create_or_update_sql_script( :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either SqlScriptResource or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2020_12_01.models.SqlScriptResource] + :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.models.SqlScriptResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] @@ -264,7 +264,7 @@ def get_sql_script( :type if_none_match: str :keyword callable cls: A custom type or function that will be passed the direct response :return: SqlScriptResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.SqlScriptResource or None + :rtype: ~azure.synapse.artifacts.models.SqlScriptResource or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.SqlScriptResource"]] @@ -299,7 +299,7 @@ def get_sql_script( if response.status_code not in [200, 304]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -348,7 +348,7 @@ def _delete_sql_script_initial( if response.status_code not in [200, 202, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -460,7 +460,7 @@ def _rename_sql_script_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_trigger_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_trigger_operations.py similarity index 98% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_trigger_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_trigger_operations.py index 4cb712e1527b..6c06c877a9c0 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_trigger_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_trigger_operations.py @@ -31,7 +31,7 @@ class TriggerOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -55,7 +55,7 @@ def get_triggers_by_workspace( :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either TriggerListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2020_12_01.models.TriggerListResponse] + :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.models.TriggerListResponse] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.TriggerListResponse"] @@ -107,7 +107,7 @@ def get_next(next_link=None): response = pipeline_response.http_response if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, model=error) @@ -165,7 +165,7 @@ def _create_or_update_trigger_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -191,7 +191,7 @@ def begin_create_or_update_trigger( :param trigger_name: The trigger name. :type trigger_name: str :param properties: Properties of the trigger. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.Trigger + :type properties: ~azure.synapse.artifacts.models.Trigger :param if_match: ETag of the trigger entity. Should only be specified for update, for which it should match existing entity or can be * for unconditional update. :type if_match: str @@ -202,7 +202,7 @@ def begin_create_or_update_trigger( :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either TriggerResource or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2020_12_01.models.TriggerResource] + :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.models.TriggerResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] @@ -266,7 +266,7 @@ def get_trigger( :type if_none_match: str :keyword callable cls: A custom type or function that will be passed the direct response :return: TriggerResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.TriggerResource or None + :rtype: ~azure.synapse.artifacts.models.TriggerResource or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.TriggerResource"]] @@ -301,7 +301,7 @@ def get_trigger( if response.status_code not in [200, 304]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -350,7 +350,7 @@ def _delete_trigger_initial( if response.status_code not in [200, 202, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -454,7 +454,7 @@ def _subscribe_trigger_to_events_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -484,7 +484,7 @@ def begin_subscribe_trigger_to_events( :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either TriggerSubscriptionOperationStatus or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2020_12_01.models.TriggerSubscriptionOperationStatus] + :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.models.TriggerSubscriptionOperationStatus] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] @@ -542,7 +542,7 @@ def get_event_subscription_status( :type trigger_name: str :keyword callable cls: A custom type or function that will be passed the direct response :return: TriggerSubscriptionOperationStatus, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.TriggerSubscriptionOperationStatus + :rtype: ~azure.synapse.artifacts.models.TriggerSubscriptionOperationStatus :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.TriggerSubscriptionOperationStatus"] @@ -575,7 +575,7 @@ def get_event_subscription_status( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = self._deserialize('TriggerSubscriptionOperationStatus', pipeline_response) @@ -622,7 +622,7 @@ def _unsubscribe_trigger_from_events_initial( if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = None @@ -652,7 +652,7 @@ def begin_unsubscribe_trigger_from_events( :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either TriggerSubscriptionOperationStatus or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2020_12_01.models.TriggerSubscriptionOperationStatus] + :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.models.TriggerSubscriptionOperationStatus] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] @@ -734,7 +734,7 @@ def _start_trigger_initial( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -838,7 +838,7 @@ def _stop_trigger_initial( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_trigger_run_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_trigger_run_operations.py similarity index 96% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_trigger_run_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_trigger_run_operations.py index 60895ba07290..a320f7eef08a 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_trigger_run_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_trigger_run_operations.py @@ -28,7 +28,7 @@ class TriggerRunOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -92,7 +92,7 @@ def rerun_trigger_instance( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -149,7 +149,7 @@ def cancel_trigger_instance( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) if cls: @@ -166,10 +166,10 @@ def query_trigger_runs_by_workspace( """Query trigger runs. :param filter_parameters: Parameters to filter the pipeline run. - :type filter_parameters: ~azure.synapse.artifacts.v2020_12_01.models.RunFilterParameters + :type filter_parameters: ~azure.synapse.artifacts.models.RunFilterParameters :keyword callable cls: A custom type or function that will be passed the direct response :return: TriggerRunsQueryResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.TriggerRunsQueryResponse + :rtype: ~azure.synapse.artifacts.models.TriggerRunsQueryResponse :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.TriggerRunsQueryResponse"] @@ -206,7 +206,7 @@ def query_trigger_runs_by_workspace( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) + error = self._deserialize.failsafe_deserialize(_models.CloudErrorAutoGenerated, response) raise HttpResponseError(response=response, model=error) deserialized = self._deserialize('TriggerRunsQueryResponse', pipeline_response) diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_workspace_git_repo_management_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_workspace_git_repo_management_operations.py similarity index 96% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_workspace_git_repo_management_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_workspace_git_repo_management_operations.py index aa79ac1c95ba..f629c472566f 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_workspace_git_repo_management_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_workspace_git_repo_management_operations.py @@ -28,7 +28,7 @@ class WorkspaceGitRepoManagementOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -53,13 +53,13 @@ def get_git_hub_access_token( """Get the GitHub access token. :param git_hub_access_token_request: - :type git_hub_access_token_request: ~azure.synapse.artifacts.v2020_12_01.models.GitHubAccessTokenRequest + :type git_hub_access_token_request: ~azure.synapse.artifacts.models.GitHubAccessTokenRequest :param client_request_id: Can provide a guid, which is helpful for debugging and to provide better customer support. :type client_request_id: str :keyword callable cls: A custom type or function that will be passed the direct response :return: GitHubAccessTokenResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.GitHubAccessTokenResponse + :rtype: ~azure.synapse.artifacts.models.GitHubAccessTokenResponse :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.GitHubAccessTokenResponse"] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_workspace_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_workspace_operations.py similarity index 96% rename from sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_workspace_operations.py rename to sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_workspace_operations.py index 2f4d607e949d..b98161b819bd 100644 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_workspace_operations.py +++ b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/operations/_workspace_operations.py @@ -28,7 +28,7 @@ class WorkspaceOperations(object): instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models + :type models: ~azure.synapse.artifacts.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. @@ -52,7 +52,7 @@ def get( :keyword callable cls: A custom type or function that will be passed the direct response :return: Workspace, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2020_12_01.models.Workspace + :rtype: ~azure.synapse.artifacts.models.Workspace :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.Workspace"] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/__init__.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/__init__.py deleted file mode 100644 index 03703237ffc5..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/__init__.py +++ /dev/null @@ -1,19 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from ._artifacts_client import ArtifactsClient -from ._version import VERSION - -__version__ = VERSION -__all__ = ['ArtifactsClient'] - -try: - from ._patch import patch_sdk # type: ignore - patch_sdk() -except ImportError: - pass diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/_artifacts_client.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/_artifacts_client.py deleted file mode 100644 index 325329bdb332..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/_artifacts_client.py +++ /dev/null @@ -1,181 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from typing import TYPE_CHECKING - -from azure.core import PipelineClient -from msrest import Deserializer, Serializer - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any - - from azure.core.credentials import TokenCredential - from azure.core.pipeline.transport import HttpRequest, HttpResponse - -from ._configuration import ArtifactsClientConfiguration -from .operations import LinkedServiceOperations -from .operations import DatasetOperations -from .operations import PipelineOperations -from .operations import PipelineRunOperations -from .operations import TriggerOperations -from .operations import TriggerRunOperations -from .operations import DataFlowOperations -from .operations import DataFlowDebugSessionOperations -from .operations import SqlScriptOperations -from .operations import SparkJobDefinitionOperations -from .operations import NotebookOperations -from .operations import NotebookOperationResultOperations -from .operations import BigDataPoolsOperations -from .operations import WorkspaceGitRepoManagementOperations -from .operations import IntegrationRuntimesOperations -from .operations import LibraryOperations -from .operations import OperationResultOperations -from .operations import OperationStatusOperations -from .operations import SqlPoolsOperations -from .operations import WorkspaceOperations -from . import models - - -class ArtifactsClient(object): - """ArtifactsClient. - - :ivar linked_service: LinkedServiceOperations operations - :vartype linked_service: azure.synapse.artifacts.v2020_12_01.operations.LinkedServiceOperations - :ivar dataset: DatasetOperations operations - :vartype dataset: azure.synapse.artifacts.v2020_12_01.operations.DatasetOperations - :ivar pipeline: PipelineOperations operations - :vartype pipeline: azure.synapse.artifacts.v2020_12_01.operations.PipelineOperations - :ivar pipeline_run: PipelineRunOperations operations - :vartype pipeline_run: azure.synapse.artifacts.v2020_12_01.operations.PipelineRunOperations - :ivar trigger: TriggerOperations operations - :vartype trigger: azure.synapse.artifacts.v2020_12_01.operations.TriggerOperations - :ivar trigger_run: TriggerRunOperations operations - :vartype trigger_run: azure.synapse.artifacts.v2020_12_01.operations.TriggerRunOperations - :ivar data_flow: DataFlowOperations operations - :vartype data_flow: azure.synapse.artifacts.v2020_12_01.operations.DataFlowOperations - :ivar data_flow_debug_session: DataFlowDebugSessionOperations operations - :vartype data_flow_debug_session: azure.synapse.artifacts.v2020_12_01.operations.DataFlowDebugSessionOperations - :ivar sql_script: SqlScriptOperations operations - :vartype sql_script: azure.synapse.artifacts.v2020_12_01.operations.SqlScriptOperations - :ivar spark_job_definition: SparkJobDefinitionOperations operations - :vartype spark_job_definition: azure.synapse.artifacts.v2020_12_01.operations.SparkJobDefinitionOperations - :ivar notebook: NotebookOperations operations - :vartype notebook: azure.synapse.artifacts.v2020_12_01.operations.NotebookOperations - :ivar notebook_operation_result: NotebookOperationResultOperations operations - :vartype notebook_operation_result: azure.synapse.artifacts.v2020_12_01.operations.NotebookOperationResultOperations - :ivar big_data_pools: BigDataPoolsOperations operations - :vartype big_data_pools: azure.synapse.artifacts.v2020_12_01.operations.BigDataPoolsOperations - :ivar workspace_git_repo_management: WorkspaceGitRepoManagementOperations operations - :vartype workspace_git_repo_management: azure.synapse.artifacts.v2020_12_01.operations.WorkspaceGitRepoManagementOperations - :ivar integration_runtimes: IntegrationRuntimesOperations operations - :vartype integration_runtimes: azure.synapse.artifacts.v2020_12_01.operations.IntegrationRuntimesOperations - :ivar library: LibraryOperations operations - :vartype library: azure.synapse.artifacts.v2020_12_01.operations.LibraryOperations - :ivar operation_result: OperationResultOperations operations - :vartype operation_result: azure.synapse.artifacts.v2020_12_01.operations.OperationResultOperations - :ivar operation_status: OperationStatusOperations operations - :vartype operation_status: azure.synapse.artifacts.v2020_12_01.operations.OperationStatusOperations - :ivar sql_pools: SqlPoolsOperations operations - :vartype sql_pools: azure.synapse.artifacts.v2020_12_01.operations.SqlPoolsOperations - :ivar workspace: WorkspaceOperations operations - :vartype workspace: azure.synapse.artifacts.v2020_12_01.operations.WorkspaceOperations - :param credential: Credential needed for the client to connect to Azure. - :type credential: ~azure.core.credentials.TokenCredential - :param endpoint: The workspace development endpoint, for example https://myworkspace.dev.azuresynapse.net. - :type endpoint: str - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - """ - - def __init__( - self, - credential, # type: "TokenCredential" - endpoint, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - base_url = '{endpoint}' - self._config = ArtifactsClientConfiguration(credential, endpoint, **kwargs) - self._client = PipelineClient(base_url=base_url, config=self._config, **kwargs) - - client_models = {k: v for k, v in models.__dict__.items() if isinstance(v, type)} - self._serialize = Serializer(client_models) - self._serialize.client_side_validation = False - self._deserialize = Deserializer(client_models) - - self.linked_service = LinkedServiceOperations( - self._client, self._config, self._serialize, self._deserialize) - self.dataset = DatasetOperations( - self._client, self._config, self._serialize, self._deserialize) - self.pipeline = PipelineOperations( - self._client, self._config, self._serialize, self._deserialize) - self.pipeline_run = PipelineRunOperations( - self._client, self._config, self._serialize, self._deserialize) - self.trigger = TriggerOperations( - self._client, self._config, self._serialize, self._deserialize) - self.trigger_run = TriggerRunOperations( - self._client, self._config, self._serialize, self._deserialize) - self.data_flow = DataFlowOperations( - self._client, self._config, self._serialize, self._deserialize) - self.data_flow_debug_session = DataFlowDebugSessionOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_script = SqlScriptOperations( - self._client, self._config, self._serialize, self._deserialize) - self.spark_job_definition = SparkJobDefinitionOperations( - self._client, self._config, self._serialize, self._deserialize) - self.notebook = NotebookOperations( - self._client, self._config, self._serialize, self._deserialize) - self.notebook_operation_result = NotebookOperationResultOperations( - self._client, self._config, self._serialize, self._deserialize) - self.big_data_pools = BigDataPoolsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.workspace_git_repo_management = WorkspaceGitRepoManagementOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtimes = IntegrationRuntimesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.library = LibraryOperations( - self._client, self._config, self._serialize, self._deserialize) - self.operation_result = OperationResultOperations( - self._client, self._config, self._serialize, self._deserialize) - self.operation_status = OperationStatusOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pools = SqlPoolsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.workspace = WorkspaceOperations( - self._client, self._config, self._serialize, self._deserialize) - - def _send_request(self, http_request, **kwargs): - # type: (HttpRequest, Any) -> HttpResponse - """Runs the network request through the client's chained policies. - - :param http_request: The network request you want to make. Required. - :type http_request: ~azure.core.pipeline.transport.HttpRequest - :keyword bool stream: Whether the response payload will be streamed. Defaults to True. - :return: The response of your network call. Does not do error handling on your response. - :rtype: ~azure.core.pipeline.transport.HttpResponse - """ - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - http_request.url = self._client.format_url(http_request.url, **path_format_arguments) - stream = kwargs.pop("stream", True) - pipeline_response = self._client._pipeline.run(http_request, stream=stream, **kwargs) - return pipeline_response.http_response - - def close(self): - # type: () -> None - self._client.close() - - def __enter__(self): - # type: () -> ArtifactsClient - self._client.__enter__() - return self - - def __exit__(self, *exc_details): - # type: (Any) -> None - self._client.__exit__(*exc_details) diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/_configuration.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/_configuration.py deleted file mode 100644 index 87d96558da96..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/_configuration.py +++ /dev/null @@ -1,70 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from typing import TYPE_CHECKING - -from azure.core.configuration import Configuration -from azure.core.pipeline import policies - -from ._version import VERSION - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any - - from azure.core.credentials import TokenCredential - - -class ArtifactsClientConfiguration(Configuration): - """Configuration for ArtifactsClient. - - Note that all parameters used to create this instance are saved as instance - attributes. - - :param credential: Credential needed for the client to connect to Azure. - :type credential: ~azure.core.credentials.TokenCredential - :param endpoint: The workspace development endpoint, for example https://myworkspace.dev.azuresynapse.net. - :type endpoint: str - """ - - def __init__( - self, - credential, # type: "TokenCredential" - endpoint, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - if credential is None: - raise ValueError("Parameter 'credential' must not be None.") - if endpoint is None: - raise ValueError("Parameter 'endpoint' must not be None.") - super(ArtifactsClientConfiguration, self).__init__(**kwargs) - - self.credential = credential - self.endpoint = endpoint - self.api_version = "2020-12-01" - self.credential_scopes = kwargs.pop('credential_scopes', ['https://dev.azuresynapse.net/.default']) - kwargs.setdefault('sdk_moniker', 'synapse-artifacts/{}'.format(VERSION)) - self._configure(**kwargs) - - def _configure( - self, - **kwargs # type: Any - ): - # type: (...) -> None - self.user_agent_policy = kwargs.get('user_agent_policy') or policies.UserAgentPolicy(**kwargs) - self.headers_policy = kwargs.get('headers_policy') or policies.HeadersPolicy(**kwargs) - self.proxy_policy = kwargs.get('proxy_policy') or policies.ProxyPolicy(**kwargs) - self.logging_policy = kwargs.get('logging_policy') or policies.NetworkTraceLoggingPolicy(**kwargs) - self.http_logging_policy = kwargs.get('http_logging_policy') or policies.HttpLoggingPolicy(**kwargs) - self.retry_policy = kwargs.get('retry_policy') or policies.RetryPolicy(**kwargs) - self.custom_hook_policy = kwargs.get('custom_hook_policy') or policies.CustomHookPolicy(**kwargs) - self.redirect_policy = kwargs.get('redirect_policy') or policies.RedirectPolicy(**kwargs) - self.authentication_policy = kwargs.get('authentication_policy') - if self.credential and not self.authentication_policy: - self.authentication_policy = policies.BearerTokenCredentialPolicy(self.credential, *self.credential_scopes, **kwargs) diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/_metadata.json b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/_metadata.json deleted file mode 100644 index dca82d86e87d..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/_metadata.json +++ /dev/null @@ -1,110 +0,0 @@ -{ - "chosen_version": "2020-12-01", - "total_api_version_list": ["2020-12-01"], - "client": { - "name": "ArtifactsClient", - "filename": "_artifacts_client", - "description": "ArtifactsClient.", - "base_url": null, - "custom_base_url": "\u0027{endpoint}\u0027", - "azure_arm": false, - "has_lro_operations": true, - "client_side_validation": false, - "sync_imports": "{\"typing\": {\"azurecore\": {\"azure.core.credentials\": [\"TokenCredential\"]}}, \"regular\": {\"azurecore\": {\"azure.profiles\": [\"KnownProfiles\", \"ProfileDefinition\"], \"azure.profiles.multiapiclient\": [\"MultiApiClientMixin\"], \"msrest\": [\"Deserializer\", \"Serializer\"], \"azure.core\": [\"PipelineClient\"]}, \"local\": {\"._configuration\": [\"ArtifactsClientConfiguration\"]}}, \"conditional\": {\"stdlib\": {\"typing\": [\"Any\", \"Optional\"]}, \"azurecore\": {\"azure.core.pipeline.transport\": [\"HttpRequest\", \"HttpResponse\"]}}}", - "async_imports": "{\"typing\": {\"azurecore\": {\"azure.core.credentials_async\": [\"AsyncTokenCredential\"]}}, \"regular\": {\"azurecore\": {\"azure.profiles\": [\"KnownProfiles\", \"ProfileDefinition\"], \"azure.profiles.multiapiclient\": [\"MultiApiClientMixin\"], \"msrest\": [\"Deserializer\", \"Serializer\"], \"azure.core\": [\"AsyncPipelineClient\"]}, \"local\": {\"._configuration\": [\"ArtifactsClientConfiguration\"]}}, \"conditional\": {\"stdlib\": {\"typing\": [\"Any\", \"Optional\"]}, \"azurecore\": {\"azure.core.pipeline.transport\": [\"AsyncHttpResponse\", \"HttpRequest\"]}}}" - }, - "global_parameters": { - "sync": { - "credential": { - "signature": "credential, # type: \"TokenCredential\"", - "description": "Credential needed for the client to connect to Azure.", - "docstring_type": "~azure.core.credentials.TokenCredential", - "required": true - }, - "endpoint": { - "signature": "endpoint, # type: str", - "description": "The workspace development endpoint, for example https://myworkspace.dev.azuresynapse.net.", - "docstring_type": "str", - "required": true - } - }, - "async": { - "credential": { - "signature": "credential: \"AsyncTokenCredential\",", - "description": "Credential needed for the client to connect to Azure.", - "docstring_type": "~azure.core.credentials_async.AsyncTokenCredential", - "required": true - }, - "endpoint": { - "signature": "endpoint: str,", - "description": "The workspace development endpoint, for example https://myworkspace.dev.azuresynapse.net.", - "docstring_type": "str", - "required": true - } - }, - "constant": { - }, - "call": "credential, endpoint", - "service_client_specific": { - "sync": { - "api_version": { - "signature": "api_version=None, # type: Optional[str]", - "description": "API version to use if no profile is provided, or if missing in profile.", - "docstring_type": "str", - "required": false - }, - "profile": { - "signature": "profile=KnownProfiles.default, # type: KnownProfiles", - "description": "A profile definition, from KnownProfiles to dict.", - "docstring_type": "azure.profiles.KnownProfiles", - "required": false - } - }, - "async": { - "api_version": { - "signature": "api_version: Optional[str] = None,", - "description": "API version to use if no profile is provided, or if missing in profile.", - "docstring_type": "str", - "required": false - }, - "profile": { - "signature": "profile: KnownProfiles = KnownProfiles.default,", - "description": "A profile definition, from KnownProfiles to dict.", - "docstring_type": "azure.profiles.KnownProfiles", - "required": false - } - } - } - }, - "config": { - "credential": true, - "credential_scopes": ["https://dev.azuresynapse.net/.default"], - "credential_default_policy_type": "BearerTokenCredentialPolicy", - "credential_default_policy_type_has_async_version": true, - "credential_key_header_name": null, - "sync_imports": "{\"regular\": {\"azurecore\": {\"azure.core.configuration\": [\"Configuration\"], \"azure.core.pipeline\": [\"policies\"]}, \"local\": {\"._version\": [\"VERSION\"]}}, \"conditional\": {\"stdlib\": {\"typing\": [\"Any\"]}}, \"typing\": {\"azurecore\": {\"azure.core.credentials\": [\"TokenCredential\"]}}}", - "async_imports": "{\"regular\": {\"azurecore\": {\"azure.core.configuration\": [\"Configuration\"], \"azure.core.pipeline\": [\"policies\"]}, \"local\": {\".._version\": [\"VERSION\"]}}, \"conditional\": {\"stdlib\": {\"typing\": [\"Any\"]}}, \"typing\": {\"azurecore\": {\"azure.core.credentials_async\": [\"AsyncTokenCredential\"]}}}" - }, - "operation_groups": { - "linked_service": "LinkedServiceOperations", - "dataset": "DatasetOperations", - "pipeline": "PipelineOperations", - "pipeline_run": "PipelineRunOperations", - "trigger": "TriggerOperations", - "trigger_run": "TriggerRunOperations", - "data_flow": "DataFlowOperations", - "data_flow_debug_session": "DataFlowDebugSessionOperations", - "sql_script": "SqlScriptOperations", - "spark_job_definition": "SparkJobDefinitionOperations", - "notebook": "NotebookOperations", - "notebook_operation_result": "NotebookOperationResultOperations", - "big_data_pools": "BigDataPoolsOperations", - "workspace_git_repo_management": "WorkspaceGitRepoManagementOperations", - "integration_runtimes": "IntegrationRuntimesOperations", - "library": "LibraryOperations", - "operation_result": "OperationResultOperations", - "operation_status": "OperationStatusOperations", - "sql_pools": "SqlPoolsOperations", - "workspace": "WorkspaceOperations" - } -} \ No newline at end of file diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/_version.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/_version.py deleted file mode 100644 index eae7c95b6fbd..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/_version.py +++ /dev/null @@ -1,9 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -VERSION = "0.1.0" diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/__init__.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/__init__.py deleted file mode 100644 index 16f882fa0a48..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/__init__.py +++ /dev/null @@ -1,10 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from ._artifacts_client import ArtifactsClient -__all__ = ['ArtifactsClient'] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/_artifacts_client.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/_artifacts_client.py deleted file mode 100644 index a526389e5707..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/_artifacts_client.py +++ /dev/null @@ -1,174 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from typing import Any, TYPE_CHECKING - -from azure.core import AsyncPipelineClient -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest -from msrest import Deserializer, Serializer - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from azure.core.credentials_async import AsyncTokenCredential - -from ._configuration import ArtifactsClientConfiguration -from .operations import LinkedServiceOperations -from .operations import DatasetOperations -from .operations import PipelineOperations -from .operations import PipelineRunOperations -from .operations import TriggerOperations -from .operations import TriggerRunOperations -from .operations import DataFlowOperations -from .operations import DataFlowDebugSessionOperations -from .operations import SqlScriptOperations -from .operations import SparkJobDefinitionOperations -from .operations import NotebookOperations -from .operations import NotebookOperationResultOperations -from .operations import BigDataPoolsOperations -from .operations import WorkspaceGitRepoManagementOperations -from .operations import IntegrationRuntimesOperations -from .operations import LibraryOperations -from .operations import OperationResultOperations -from .operations import OperationStatusOperations -from .operations import SqlPoolsOperations -from .operations import WorkspaceOperations -from .. import models - - -class ArtifactsClient(object): - """ArtifactsClient. - - :ivar linked_service: LinkedServiceOperations operations - :vartype linked_service: azure.synapse.artifacts.v2020_12_01.aio.operations.LinkedServiceOperations - :ivar dataset: DatasetOperations operations - :vartype dataset: azure.synapse.artifacts.v2020_12_01.aio.operations.DatasetOperations - :ivar pipeline: PipelineOperations operations - :vartype pipeline: azure.synapse.artifacts.v2020_12_01.aio.operations.PipelineOperations - :ivar pipeline_run: PipelineRunOperations operations - :vartype pipeline_run: azure.synapse.artifacts.v2020_12_01.aio.operations.PipelineRunOperations - :ivar trigger: TriggerOperations operations - :vartype trigger: azure.synapse.artifacts.v2020_12_01.aio.operations.TriggerOperations - :ivar trigger_run: TriggerRunOperations operations - :vartype trigger_run: azure.synapse.artifacts.v2020_12_01.aio.operations.TriggerRunOperations - :ivar data_flow: DataFlowOperations operations - :vartype data_flow: azure.synapse.artifacts.v2020_12_01.aio.operations.DataFlowOperations - :ivar data_flow_debug_session: DataFlowDebugSessionOperations operations - :vartype data_flow_debug_session: azure.synapse.artifacts.v2020_12_01.aio.operations.DataFlowDebugSessionOperations - :ivar sql_script: SqlScriptOperations operations - :vartype sql_script: azure.synapse.artifacts.v2020_12_01.aio.operations.SqlScriptOperations - :ivar spark_job_definition: SparkJobDefinitionOperations operations - :vartype spark_job_definition: azure.synapse.artifacts.v2020_12_01.aio.operations.SparkJobDefinitionOperations - :ivar notebook: NotebookOperations operations - :vartype notebook: azure.synapse.artifacts.v2020_12_01.aio.operations.NotebookOperations - :ivar notebook_operation_result: NotebookOperationResultOperations operations - :vartype notebook_operation_result: azure.synapse.artifacts.v2020_12_01.aio.operations.NotebookOperationResultOperations - :ivar big_data_pools: BigDataPoolsOperations operations - :vartype big_data_pools: azure.synapse.artifacts.v2020_12_01.aio.operations.BigDataPoolsOperations - :ivar workspace_git_repo_management: WorkspaceGitRepoManagementOperations operations - :vartype workspace_git_repo_management: azure.synapse.artifacts.v2020_12_01.aio.operations.WorkspaceGitRepoManagementOperations - :ivar integration_runtimes: IntegrationRuntimesOperations operations - :vartype integration_runtimes: azure.synapse.artifacts.v2020_12_01.aio.operations.IntegrationRuntimesOperations - :ivar library: LibraryOperations operations - :vartype library: azure.synapse.artifacts.v2020_12_01.aio.operations.LibraryOperations - :ivar operation_result: OperationResultOperations operations - :vartype operation_result: azure.synapse.artifacts.v2020_12_01.aio.operations.OperationResultOperations - :ivar operation_status: OperationStatusOperations operations - :vartype operation_status: azure.synapse.artifacts.v2020_12_01.aio.operations.OperationStatusOperations - :ivar sql_pools: SqlPoolsOperations operations - :vartype sql_pools: azure.synapse.artifacts.v2020_12_01.aio.operations.SqlPoolsOperations - :ivar workspace: WorkspaceOperations operations - :vartype workspace: azure.synapse.artifacts.v2020_12_01.aio.operations.WorkspaceOperations - :param credential: Credential needed for the client to connect to Azure. - :type credential: ~azure.core.credentials_async.AsyncTokenCredential - :param endpoint: The workspace development endpoint, for example https://myworkspace.dev.azuresynapse.net. - :type endpoint: str - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - """ - - def __init__( - self, - credential: "AsyncTokenCredential", - endpoint: str, - **kwargs: Any - ) -> None: - base_url = '{endpoint}' - self._config = ArtifactsClientConfiguration(credential, endpoint, **kwargs) - self._client = AsyncPipelineClient(base_url=base_url, config=self._config, **kwargs) - - client_models = {k: v for k, v in models.__dict__.items() if isinstance(v, type)} - self._serialize = Serializer(client_models) - self._serialize.client_side_validation = False - self._deserialize = Deserializer(client_models) - - self.linked_service = LinkedServiceOperations( - self._client, self._config, self._serialize, self._deserialize) - self.dataset = DatasetOperations( - self._client, self._config, self._serialize, self._deserialize) - self.pipeline = PipelineOperations( - self._client, self._config, self._serialize, self._deserialize) - self.pipeline_run = PipelineRunOperations( - self._client, self._config, self._serialize, self._deserialize) - self.trigger = TriggerOperations( - self._client, self._config, self._serialize, self._deserialize) - self.trigger_run = TriggerRunOperations( - self._client, self._config, self._serialize, self._deserialize) - self.data_flow = DataFlowOperations( - self._client, self._config, self._serialize, self._deserialize) - self.data_flow_debug_session = DataFlowDebugSessionOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_script = SqlScriptOperations( - self._client, self._config, self._serialize, self._deserialize) - self.spark_job_definition = SparkJobDefinitionOperations( - self._client, self._config, self._serialize, self._deserialize) - self.notebook = NotebookOperations( - self._client, self._config, self._serialize, self._deserialize) - self.notebook_operation_result = NotebookOperationResultOperations( - self._client, self._config, self._serialize, self._deserialize) - self.big_data_pools = BigDataPoolsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.workspace_git_repo_management = WorkspaceGitRepoManagementOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtimes = IntegrationRuntimesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.library = LibraryOperations( - self._client, self._config, self._serialize, self._deserialize) - self.operation_result = OperationResultOperations( - self._client, self._config, self._serialize, self._deserialize) - self.operation_status = OperationStatusOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pools = SqlPoolsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.workspace = WorkspaceOperations( - self._client, self._config, self._serialize, self._deserialize) - - async def _send_request(self, http_request: HttpRequest, **kwargs: Any) -> AsyncHttpResponse: - """Runs the network request through the client's chained policies. - - :param http_request: The network request you want to make. Required. - :type http_request: ~azure.core.pipeline.transport.HttpRequest - :keyword bool stream: Whether the response payload will be streamed. Defaults to True. - :return: The response of your network call. Does not do error handling on your response. - :rtype: ~azure.core.pipeline.transport.AsyncHttpResponse - """ - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - http_request.url = self._client.format_url(http_request.url, **path_format_arguments) - stream = kwargs.pop("stream", True) - pipeline_response = await self._client._pipeline.run(http_request, stream=stream, **kwargs) - return pipeline_response.http_response - - async def close(self) -> None: - await self._client.close() - - async def __aenter__(self) -> "ArtifactsClient": - await self._client.__aenter__() - return self - - async def __aexit__(self, *exc_details) -> None: - await self._client.__aexit__(*exc_details) diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/_configuration.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/_configuration.py deleted file mode 100644 index edf128be9302..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/_configuration.py +++ /dev/null @@ -1,66 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from typing import Any, TYPE_CHECKING - -from azure.core.configuration import Configuration -from azure.core.pipeline import policies - -from .._version import VERSION - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from azure.core.credentials_async import AsyncTokenCredential - - -class ArtifactsClientConfiguration(Configuration): - """Configuration for ArtifactsClient. - - Note that all parameters used to create this instance are saved as instance - attributes. - - :param credential: Credential needed for the client to connect to Azure. - :type credential: ~azure.core.credentials_async.AsyncTokenCredential - :param endpoint: The workspace development endpoint, for example https://myworkspace.dev.azuresynapse.net. - :type endpoint: str - """ - - def __init__( - self, - credential: "AsyncTokenCredential", - endpoint: str, - **kwargs: Any - ) -> None: - if credential is None: - raise ValueError("Parameter 'credential' must not be None.") - if endpoint is None: - raise ValueError("Parameter 'endpoint' must not be None.") - super(ArtifactsClientConfiguration, self).__init__(**kwargs) - - self.credential = credential - self.endpoint = endpoint - self.api_version = "2020-12-01" - self.credential_scopes = kwargs.pop('credential_scopes', ['https://dev.azuresynapse.net/.default']) - kwargs.setdefault('sdk_moniker', 'synapse-artifacts/{}'.format(VERSION)) - self._configure(**kwargs) - - def _configure( - self, - **kwargs: Any - ) -> None: - self.user_agent_policy = kwargs.get('user_agent_policy') or policies.UserAgentPolicy(**kwargs) - self.headers_policy = kwargs.get('headers_policy') or policies.HeadersPolicy(**kwargs) - self.proxy_policy = kwargs.get('proxy_policy') or policies.ProxyPolicy(**kwargs) - self.logging_policy = kwargs.get('logging_policy') or policies.NetworkTraceLoggingPolicy(**kwargs) - self.http_logging_policy = kwargs.get('http_logging_policy') or policies.HttpLoggingPolicy(**kwargs) - self.retry_policy = kwargs.get('retry_policy') or policies.AsyncRetryPolicy(**kwargs) - self.custom_hook_policy = kwargs.get('custom_hook_policy') or policies.CustomHookPolicy(**kwargs) - self.redirect_policy = kwargs.get('redirect_policy') or policies.AsyncRedirectPolicy(**kwargs) - self.authentication_policy = kwargs.get('authentication_policy') - if self.credential and not self.authentication_policy: - self.authentication_policy = policies.AsyncBearerTokenCredentialPolicy(self.credential, *self.credential_scopes, **kwargs) diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/__init__.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/__init__.py deleted file mode 100644 index ddf9d04482ce..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/__init__.py +++ /dev/null @@ -1,51 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from ._linked_service_operations import LinkedServiceOperations -from ._dataset_operations import DatasetOperations -from ._pipeline_operations import PipelineOperations -from ._pipeline_run_operations import PipelineRunOperations -from ._trigger_operations import TriggerOperations -from ._trigger_run_operations import TriggerRunOperations -from ._data_flow_operations import DataFlowOperations -from ._data_flow_debug_session_operations import DataFlowDebugSessionOperations -from ._sql_script_operations import SqlScriptOperations -from ._spark_job_definition_operations import SparkJobDefinitionOperations -from ._notebook_operations import NotebookOperations -from ._notebook_operation_result_operations import NotebookOperationResultOperations -from ._big_data_pools_operations import BigDataPoolsOperations -from ._workspace_git_repo_management_operations import WorkspaceGitRepoManagementOperations -from ._integration_runtimes_operations import IntegrationRuntimesOperations -from ._library_operations import LibraryOperations -from ._operation_result_operations import OperationResultOperations -from ._operation_status_operations import OperationStatusOperations -from ._sql_pools_operations import SqlPoolsOperations -from ._workspace_operations import WorkspaceOperations - -__all__ = [ - 'LinkedServiceOperations', - 'DatasetOperations', - 'PipelineOperations', - 'PipelineRunOperations', - 'TriggerOperations', - 'TriggerRunOperations', - 'DataFlowOperations', - 'DataFlowDebugSessionOperations', - 'SqlScriptOperations', - 'SparkJobDefinitionOperations', - 'NotebookOperations', - 'NotebookOperationResultOperations', - 'BigDataPoolsOperations', - 'WorkspaceGitRepoManagementOperations', - 'IntegrationRuntimesOperations', - 'LibraryOperations', - 'OperationResultOperations', - 'OperationStatusOperations', - 'SqlPoolsOperations', - 'WorkspaceOperations', -] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_operation_result_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_operation_result_operations.py deleted file mode 100644 index 212f6b8eb57c..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_operation_result_operations.py +++ /dev/null @@ -1,92 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, Callable, Dict, Generic, Optional, TypeVar -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class OperationResultOperations: - """OperationResultOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - async def get( - self, - operation_id: str, - **kwargs: Any - ) -> None: - """Get operation result. - - :param operation_id: Operation ID. - :type operation_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2020-12-01" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'operationId': self._serialize.url("operation_id", operation_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 201, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - get.metadata = {'url': '/operationResults/{operationId}'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_operation_status_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_operation_status_operations.py deleted file mode 100644 index 06ba0fa196dd..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/aio/operations/_operation_status_operations.py +++ /dev/null @@ -1,92 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, Callable, Dict, Generic, Optional, TypeVar -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class OperationStatusOperations: - """OperationStatusOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - async def get( - self, - operation_id: str, - **kwargs: Any - ) -> None: - """Get operation status. - - :param operation_id: Operation ID. - :type operation_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2020-12-01" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'operationId': self._serialize.url("operation_id", operation_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 201, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - get.metadata = {'url': '/operationStatuses/{operationId}'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/models/__init__.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/models/__init__.py deleted file mode 100644 index b56306f02287..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/models/__init__.py +++ /dev/null @@ -1,2187 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -try: - from ._models_py3 import Activity - from ._models_py3 import ActivityDependency - from ._models_py3 import ActivityPolicy - from ._models_py3 import ActivityRun - from ._models_py3 import ActivityRunsQueryResponse - from ._models_py3 import AddDataFlowToDebugSessionResponse - from ._models_py3 import AdditionalColumns - from ._models_py3 import AmazonMWSLinkedService - from ._models_py3 import AmazonMWSObjectDataset - from ._models_py3 import AmazonMWSSource - from ._models_py3 import AmazonRedshiftLinkedService - from ._models_py3 import AmazonRedshiftSource - from ._models_py3 import AmazonRedshiftTableDataset - from ._models_py3 import AmazonS3Dataset - from ._models_py3 import AmazonS3LinkedService - from ._models_py3 import AmazonS3Location - from ._models_py3 import AmazonS3ReadSettings - from ._models_py3 import AppendVariableActivity - from ._models_py3 import ArtifactRenameRequest - from ._models_py3 import AutoPauseProperties - from ._models_py3 import AutoScaleProperties - from ._models_py3 import AvroDataset - from ._models_py3 import AvroFormat - from ._models_py3 import AvroSink - from ._models_py3 import AvroSource - from ._models_py3 import AvroWriteSettings - from ._models_py3 import AzureBatchLinkedService - from ._models_py3 import AzureBlobDataset - from ._models_py3 import AzureBlobFSDataset - from ._models_py3 import AzureBlobFSLinkedService - from ._models_py3 import AzureBlobFSLocation - from ._models_py3 import AzureBlobFSReadSettings - from ._models_py3 import AzureBlobFSSink - from ._models_py3 import AzureBlobFSSource - from ._models_py3 import AzureBlobFSWriteSettings - from ._models_py3 import AzureBlobStorageLinkedService - from ._models_py3 import AzureBlobStorageLocation - from ._models_py3 import AzureBlobStorageReadSettings - from ._models_py3 import AzureBlobStorageWriteSettings - from ._models_py3 import AzureDataExplorerCommandActivity - from ._models_py3 import AzureDataExplorerLinkedService - from ._models_py3 import AzureDataExplorerSink - from ._models_py3 import AzureDataExplorerSource - from ._models_py3 import AzureDataExplorerTableDataset - from ._models_py3 import AzureDataLakeAnalyticsLinkedService - from ._models_py3 import AzureDataLakeStoreDataset - from ._models_py3 import AzureDataLakeStoreLinkedService - from ._models_py3 import AzureDataLakeStoreLocation - from ._models_py3 import AzureDataLakeStoreReadSettings - from ._models_py3 import AzureDataLakeStoreSink - from ._models_py3 import AzureDataLakeStoreSource - from ._models_py3 import AzureDataLakeStoreWriteSettings - from ._models_py3 import AzureDatabricksDeltaLakeDataset - from ._models_py3 import AzureDatabricksDeltaLakeExportCommand - from ._models_py3 import AzureDatabricksDeltaLakeImportCommand - from ._models_py3 import AzureDatabricksDeltaLakeLinkedService - from ._models_py3 import AzureDatabricksDeltaLakeSink - from ._models_py3 import AzureDatabricksDeltaLakeSource - from ._models_py3 import AzureDatabricksLinkedService - from ._models_py3 import AzureEntityResource - from ._models_py3 import AzureFileStorageLinkedService - from ._models_py3 import AzureFileStorageLocation - from ._models_py3 import AzureFileStorageReadSettings - from ._models_py3 import AzureFileStorageWriteSettings - from ._models_py3 import AzureFunctionActivity - from ._models_py3 import AzureFunctionLinkedService - from ._models_py3 import AzureKeyVaultLinkedService - from ._models_py3 import AzureKeyVaultSecretReference - from ._models_py3 import AzureMLBatchExecutionActivity - from ._models_py3 import AzureMLExecutePipelineActivity - from ._models_py3 import AzureMLLinkedService - from ._models_py3 import AzureMLServiceLinkedService - from ._models_py3 import AzureMLUpdateResourceActivity - from ._models_py3 import AzureMLWebServiceFile - from ._models_py3 import AzureMariaDBLinkedService - from ._models_py3 import AzureMariaDBSource - from ._models_py3 import AzureMariaDBTableDataset - from ._models_py3 import AzureMySqlLinkedService - from ._models_py3 import AzureMySqlSink - from ._models_py3 import AzureMySqlSource - from ._models_py3 import AzureMySqlTableDataset - from ._models_py3 import AzurePostgreSqlLinkedService - from ._models_py3 import AzurePostgreSqlSink - from ._models_py3 import AzurePostgreSqlSource - from ._models_py3 import AzurePostgreSqlTableDataset - from ._models_py3 import AzureQueueSink - from ._models_py3 import AzureSearchIndexDataset - from ._models_py3 import AzureSearchIndexSink - from ._models_py3 import AzureSearchLinkedService - from ._models_py3 import AzureSqlDWLinkedService - from ._models_py3 import AzureSqlDWTableDataset - from ._models_py3 import AzureSqlDatabaseLinkedService - from ._models_py3 import AzureSqlMILinkedService - from ._models_py3 import AzureSqlMITableDataset - from ._models_py3 import AzureSqlSink - from ._models_py3 import AzureSqlSource - from ._models_py3 import AzureSqlTableDataset - from ._models_py3 import AzureStorageLinkedService - from ._models_py3 import AzureTableDataset - from ._models_py3 import AzureTableSink - from ._models_py3 import AzureTableSource - from ._models_py3 import AzureTableStorageLinkedService - from ._models_py3 import BigDataPoolReference - from ._models_py3 import BigDataPoolResourceInfo - from ._models_py3 import BigDataPoolResourceInfoListResult - from ._models_py3 import BinaryDataset - from ._models_py3 import BinaryReadSettings - from ._models_py3 import BinarySink - from ._models_py3 import BinarySource - from ._models_py3 import BlobEventsTrigger - from ._models_py3 import BlobSink - from ._models_py3 import BlobSource - from ._models_py3 import BlobTrigger - from ._models_py3 import CassandraLinkedService - from ._models_py3 import CassandraSource - from ._models_py3 import CassandraTableDataset - from ._models_py3 import ChainingTrigger - from ._models_py3 import CloudError - from ._models_py3 import CommonDataServiceForAppsEntityDataset - from ._models_py3 import CommonDataServiceForAppsLinkedService - from ._models_py3 import CommonDataServiceForAppsSink - from ._models_py3 import CommonDataServiceForAppsSource - from ._models_py3 import CompressionReadSettings - from ._models_py3 import ConcurLinkedService - from ._models_py3 import ConcurObjectDataset - from ._models_py3 import ConcurSource - from ._models_py3 import ControlActivity - from ._models_py3 import CopyActivity - from ._models_py3 import CopyActivityLogSettings - from ._models_py3 import CopySink - from ._models_py3 import CopySource - from ._models_py3 import CopyTranslator - from ._models_py3 import CosmosDbLinkedService - from ._models_py3 import CosmosDbMongoDbApiCollectionDataset - from ._models_py3 import CosmosDbMongoDbApiLinkedService - from ._models_py3 import CosmosDbMongoDbApiSink - from ._models_py3 import CosmosDbMongoDbApiSource - from ._models_py3 import CosmosDbSqlApiCollectionDataset - from ._models_py3 import CosmosDbSqlApiSink - from ._models_py3 import CosmosDbSqlApiSource - from ._models_py3 import CouchbaseLinkedService - from ._models_py3 import CouchbaseSource - from ._models_py3 import CouchbaseTableDataset - from ._models_py3 import CreateDataFlowDebugSessionRequest - from ._models_py3 import CreateDataFlowDebugSessionResponse - from ._models_py3 import CreateRunResponse - from ._models_py3 import CustomActivity - from ._models_py3 import CustomActivityReferenceObject - from ._models_py3 import CustomDataSourceLinkedService - from ._models_py3 import CustomDataset - from ._models_py3 import CustomEventsTrigger - from ._models_py3 import CustomSetupBase - from ._models_py3 import CustomerManagedKeyDetails - from ._models_py3 import DWCopyCommandDefaultValue - from ._models_py3 import DWCopyCommandSettings - from ._models_py3 import DataFlow - from ._models_py3 import DataFlowDebugCommandRequest - from ._models_py3 import DataFlowDebugCommandResponse - from ._models_py3 import DataFlowDebugPackage - from ._models_py3 import DataFlowDebugPackageDebugSettings - from ._models_py3 import DataFlowDebugPreviewDataRequest - from ._models_py3 import DataFlowDebugQueryResponse - from ._models_py3 import DataFlowDebugResource - from ._models_py3 import DataFlowDebugResultResponse - from ._models_py3 import DataFlowDebugSessionInfo - from ._models_py3 import DataFlowDebugStatisticsRequest - from ._models_py3 import DataFlowFolder - from ._models_py3 import DataFlowListResponse - from ._models_py3 import DataFlowReference - from ._models_py3 import DataFlowResource - from ._models_py3 import DataFlowSink - from ._models_py3 import DataFlowSource - from ._models_py3 import DataFlowSourceSetting - from ._models_py3 import DataFlowStagingInfo - from ._models_py3 import DataLakeAnalyticsUSQLActivity - from ._models_py3 import DataLakeStorageAccountDetails - from ._models_py3 import DatabricksNotebookActivity - from ._models_py3 import DatabricksSparkJarActivity - from ._models_py3 import DatabricksSparkPythonActivity - from ._models_py3 import Dataset - from ._models_py3 import DatasetBZip2Compression - from ._models_py3 import DatasetCompression - from ._models_py3 import DatasetDataElement - from ._models_py3 import DatasetDebugResource - from ._models_py3 import DatasetDeflateCompression - from ._models_py3 import DatasetFolder - from ._models_py3 import DatasetGZipCompression - from ._models_py3 import DatasetListResponse - from ._models_py3 import DatasetLocation - from ._models_py3 import DatasetReference - from ._models_py3 import DatasetResource - from ._models_py3 import DatasetSchemaDataElement - from ._models_py3 import DatasetStorageFormat - from ._models_py3 import DatasetTarCompression - from ._models_py3 import DatasetTarGZipCompression - from ._models_py3 import DatasetZipDeflateCompression - from ._models_py3 import Db2LinkedService - from ._models_py3 import Db2Source - from ._models_py3 import Db2TableDataset - from ._models_py3 import DeleteActivity - from ._models_py3 import DeleteDataFlowDebugSessionRequest - from ._models_py3 import DelimitedTextDataset - from ._models_py3 import DelimitedTextReadSettings - from ._models_py3 import DelimitedTextSink - from ._models_py3 import DelimitedTextSource - from ._models_py3 import DelimitedTextWriteSettings - from ._models_py3 import DependencyReference - from ._models_py3 import DistcpSettings - from ._models_py3 import DocumentDbCollectionDataset - from ._models_py3 import DocumentDbCollectionSink - from ._models_py3 import DocumentDbCollectionSource - from ._models_py3 import DrillLinkedService - from ._models_py3 import DrillSource - from ._models_py3 import DrillTableDataset - from ._models_py3 import DynamicExecutorAllocation - from ._models_py3 import DynamicsAXLinkedService - from ._models_py3 import DynamicsAXResourceDataset - from ._models_py3 import DynamicsAXSource - from ._models_py3 import DynamicsCrmEntityDataset - from ._models_py3 import DynamicsCrmLinkedService - from ._models_py3 import DynamicsCrmSink - from ._models_py3 import DynamicsCrmSource - from ._models_py3 import DynamicsEntityDataset - from ._models_py3 import DynamicsLinkedService - from ._models_py3 import DynamicsSink - from ._models_py3 import DynamicsSource - from ._models_py3 import EloquaLinkedService - from ._models_py3 import EloquaObjectDataset - from ._models_py3 import EloquaSource - from ._models_py3 import EncryptionDetails - from ._models_py3 import EntityReference - from ._models_py3 import ErrorAdditionalInfo - from ._models_py3 import ErrorContract - from ._models_py3 import ErrorResponse - from ._models_py3 import EvaluateDataFlowExpressionRequest - from ._models_py3 import ExcelDataset - from ._models_py3 import ExcelSource - from ._models_py3 import ExecuteDataFlowActivity - from ._models_py3 import ExecuteDataFlowActivityTypePropertiesCompute - from ._models_py3 import ExecutePipelineActivity - from ._models_py3 import ExecuteSSISPackageActivity - from ._models_py3 import ExecutionActivity - from ._models_py3 import ExportSettings - from ._models_py3 import ExposureControlRequest - from ._models_py3 import ExposureControlResponse - from ._models_py3 import Expression - from ._models_py3 import FileServerLinkedService - from ._models_py3 import FileServerLocation - from ._models_py3 import FileServerReadSettings - from ._models_py3 import FileServerWriteSettings - from ._models_py3 import FileShareDataset - from ._models_py3 import FileSystemSink - from ._models_py3 import FileSystemSource - from ._models_py3 import FilterActivity - from ._models_py3 import ForEachActivity - from ._models_py3 import FormatReadSettings - from ._models_py3 import FormatWriteSettings - from ._models_py3 import FtpReadSettings - from ._models_py3 import FtpServerLinkedService - from ._models_py3 import FtpServerLocation - from ._models_py3 import GetMetadataActivity - from ._models_py3 import GetSsisObjectMetadataRequest - from ._models_py3 import GitHubAccessTokenRequest - from ._models_py3 import GitHubAccessTokenResponse - from ._models_py3 import GitHubClientSecret - from ._models_py3 import GoogleAdWordsLinkedService - from ._models_py3 import GoogleAdWordsObjectDataset - from ._models_py3 import GoogleAdWordsSource - from ._models_py3 import GoogleBigQueryLinkedService - from ._models_py3 import GoogleBigQueryObjectDataset - from ._models_py3 import GoogleBigQuerySource - from ._models_py3 import GoogleCloudStorageLinkedService - from ._models_py3 import GoogleCloudStorageLocation - from ._models_py3 import GoogleCloudStorageReadSettings - from ._models_py3 import GreenplumLinkedService - from ._models_py3 import GreenplumSource - from ._models_py3 import GreenplumTableDataset - from ._models_py3 import HBaseLinkedService - from ._models_py3 import HBaseObjectDataset - from ._models_py3 import HBaseSource - from ._models_py3 import HDInsightHiveActivity - from ._models_py3 import HDInsightLinkedService - from ._models_py3 import HDInsightMapReduceActivity - from ._models_py3 import HDInsightOnDemandLinkedService - from ._models_py3 import HDInsightPigActivity - from ._models_py3 import HDInsightSparkActivity - from ._models_py3 import HDInsightStreamingActivity - from ._models_py3 import HdfsLinkedService - from ._models_py3 import HdfsLocation - from ._models_py3 import HdfsReadSettings - from ._models_py3 import HdfsSource - from ._models_py3 import HiveLinkedService - from ._models_py3 import HiveObjectDataset - from ._models_py3 import HiveSource - from ._models_py3 import HttpDataset - from ._models_py3 import HttpLinkedService - from ._models_py3 import HttpReadSettings - from ._models_py3 import HttpServerLocation - from ._models_py3 import HttpSource - from ._models_py3 import HubspotLinkedService - from ._models_py3 import HubspotObjectDataset - from ._models_py3 import HubspotSource - from ._models_py3 import IfConditionActivity - from ._models_py3 import ImpalaLinkedService - from ._models_py3 import ImpalaObjectDataset - from ._models_py3 import ImpalaSource - from ._models_py3 import ImportSettings - from ._models_py3 import InformixLinkedService - from ._models_py3 import InformixSink - from ._models_py3 import InformixSource - from ._models_py3 import InformixTableDataset - from ._models_py3 import IntegrationRuntime - from ._models_py3 import IntegrationRuntimeComputeProperties - from ._models_py3 import IntegrationRuntimeCustomSetupScriptProperties - from ._models_py3 import IntegrationRuntimeDataFlowProperties - from ._models_py3 import IntegrationRuntimeDataProxyProperties - from ._models_py3 import IntegrationRuntimeListResponse - from ._models_py3 import IntegrationRuntimeReference - from ._models_py3 import IntegrationRuntimeResource - from ._models_py3 import IntegrationRuntimeSsisCatalogInfo - from ._models_py3 import IntegrationRuntimeSsisProperties - from ._models_py3 import IntegrationRuntimeVNetProperties - from ._models_py3 import JiraLinkedService - from ._models_py3 import JiraObjectDataset - from ._models_py3 import JiraSource - from ._models_py3 import JsonDataset - from ._models_py3 import JsonFormat - from ._models_py3 import JsonReadSettings - from ._models_py3 import JsonSink - from ._models_py3 import JsonSource - from ._models_py3 import JsonWriteSettings - from ._models_py3 import LibraryInfo - from ._models_py3 import LibraryListResponse - from ._models_py3 import LibraryRequirements - from ._models_py3 import LibraryResource - from ._models_py3 import LibraryResourceInfo - from ._models_py3 import LibraryResourceProperties - from ._models_py3 import LinkedIntegrationRuntimeKeyAuthorization - from ._models_py3 import LinkedIntegrationRuntimeRbacAuthorization - from ._models_py3 import LinkedIntegrationRuntimeType - from ._models_py3 import LinkedService - from ._models_py3 import LinkedServiceDebugResource - from ._models_py3 import LinkedServiceListResponse - from ._models_py3 import LinkedServiceReference - from ._models_py3 import LinkedServiceResource - from ._models_py3 import LogLocationSettings - from ._models_py3 import LogSettings - from ._models_py3 import LogStorageSettings - from ._models_py3 import LookupActivity - from ._models_py3 import MagentoLinkedService - from ._models_py3 import MagentoObjectDataset - from ._models_py3 import MagentoSource - from ._models_py3 import ManagedIdentity - from ._models_py3 import ManagedIntegrationRuntime - from ._models_py3 import ManagedVirtualNetworkReference - from ._models_py3 import ManagedVirtualNetworkSettings - from ._models_py3 import MappingDataFlow - from ._models_py3 import MariaDBLinkedService - from ._models_py3 import MariaDBSource - from ._models_py3 import MariaDBTableDataset - from ._models_py3 import MarketoLinkedService - from ._models_py3 import MarketoObjectDataset - from ._models_py3 import MarketoSource - from ._models_py3 import MicrosoftAccessLinkedService - from ._models_py3 import MicrosoftAccessSink - from ._models_py3 import MicrosoftAccessSource - from ._models_py3 import MicrosoftAccessTableDataset - from ._models_py3 import MongoDbAtlasCollectionDataset - from ._models_py3 import MongoDbAtlasLinkedService - from ._models_py3 import MongoDbAtlasSource - from ._models_py3 import MongoDbCollectionDataset - from ._models_py3 import MongoDbCursorMethodsProperties - from ._models_py3 import MongoDbLinkedService - from ._models_py3 import MongoDbSource - from ._models_py3 import MongoDbV2CollectionDataset - from ._models_py3 import MongoDbV2LinkedService - from ._models_py3 import MongoDbV2Source - from ._models_py3 import MultiplePipelineTrigger - from ._models_py3 import MySqlLinkedService - from ._models_py3 import MySqlSource - from ._models_py3 import MySqlTableDataset - from ._models_py3 import NetezzaLinkedService - from ._models_py3 import NetezzaPartitionSettings - from ._models_py3 import NetezzaSource - from ._models_py3 import NetezzaTableDataset - from ._models_py3 import Notebook - from ._models_py3 import NotebookCell - from ._models_py3 import NotebookCellOutputItem - from ._models_py3 import NotebookKernelSpec - from ._models_py3 import NotebookLanguageInfo - from ._models_py3 import NotebookListResponse - from ._models_py3 import NotebookMetadata - from ._models_py3 import NotebookResource - from ._models_py3 import NotebookSessionProperties - from ._models_py3 import ODataLinkedService - from ._models_py3 import ODataResourceDataset - from ._models_py3 import ODataSource - from ._models_py3 import OdbcLinkedService - from ._models_py3 import OdbcSink - from ._models_py3 import OdbcSource - from ._models_py3 import OdbcTableDataset - from ._models_py3 import Office365Dataset - from ._models_py3 import Office365LinkedService - from ._models_py3 import Office365Source - from ._models_py3 import OperationResult - from ._models_py3 import OracleLinkedService - from ._models_py3 import OraclePartitionSettings - from ._models_py3 import OracleServiceCloudLinkedService - from ._models_py3 import OracleServiceCloudObjectDataset - from ._models_py3 import OracleServiceCloudSource - from ._models_py3 import OracleSink - from ._models_py3 import OracleSource - from ._models_py3 import OracleTableDataset - from ._models_py3 import OrcDataset - from ._models_py3 import OrcFormat - from ._models_py3 import OrcSink - from ._models_py3 import OrcSource - from ._models_py3 import OrcWriteSettings - from ._models_py3 import ParameterSpecification - from ._models_py3 import ParquetDataset - from ._models_py3 import ParquetFormat - from ._models_py3 import ParquetSink - from ._models_py3 import ParquetSource - from ._models_py3 import ParquetWriteSettings - from ._models_py3 import PaypalLinkedService - from ._models_py3 import PaypalObjectDataset - from ._models_py3 import PaypalSource - from ._models_py3 import PhoenixLinkedService - from ._models_py3 import PhoenixObjectDataset - from ._models_py3 import PhoenixSource - from ._models_py3 import PipelineFolder - from ._models_py3 import PipelineListResponse - from ._models_py3 import PipelineReference - from ._models_py3 import PipelineResource - from ._models_py3 import PipelineRun - from ._models_py3 import PipelineRunInvokedBy - from ._models_py3 import PipelineRunsQueryResponse - from ._models_py3 import PolybaseSettings - from ._models_py3 import PostgreSqlLinkedService - from ._models_py3 import PostgreSqlSource - from ._models_py3 import PostgreSqlTableDataset - from ._models_py3 import PrestoLinkedService - from ._models_py3 import PrestoObjectDataset - from ._models_py3 import PrestoSource - from ._models_py3 import PrivateEndpoint - from ._models_py3 import PrivateEndpointConnection - from ._models_py3 import PrivateLinkServiceConnectionState - from ._models_py3 import ProxyResource - from ._models_py3 import PurviewConfiguration - from ._models_py3 import QueryDataFlowDebugSessionsResponse - from ._models_py3 import QuickBooksLinkedService - from ._models_py3 import QuickBooksObjectDataset - from ._models_py3 import QuickBooksSource - from ._models_py3 import RecurrenceSchedule - from ._models_py3 import RecurrenceScheduleOccurrence - from ._models_py3 import RedirectIncompatibleRowSettings - from ._models_py3 import RedshiftUnloadSettings - from ._models_py3 import RelationalSource - from ._models_py3 import RelationalTableDataset - from ._models_py3 import RerunTriggerListResponse - from ._models_py3 import RerunTriggerResource - from ._models_py3 import RerunTumblingWindowTrigger - from ._models_py3 import RerunTumblingWindowTriggerActionParameters - from ._models_py3 import Resource - from ._models_py3 import ResponsysLinkedService - from ._models_py3 import ResponsysObjectDataset - from ._models_py3 import ResponsysSource - from ._models_py3 import RestResourceDataset - from ._models_py3 import RestServiceLinkedService - from ._models_py3 import RestSink - from ._models_py3 import RestSource - from ._models_py3 import RetryPolicy - from ._models_py3 import RunFilterParameters - from ._models_py3 import RunQueryFilter - from ._models_py3 import RunQueryOrderBy - from ._models_py3 import SSISAccessCredential - from ._models_py3 import SSISChildPackage - from ._models_py3 import SSISExecutionCredential - from ._models_py3 import SSISExecutionParameter - from ._models_py3 import SSISLogLocation - from ._models_py3 import SSISPackageLocation - from ._models_py3 import SSISPropertyOverride - from ._models_py3 import SalesforceLinkedService - from ._models_py3 import SalesforceMarketingCloudLinkedService - from ._models_py3 import SalesforceMarketingCloudObjectDataset - from ._models_py3 import SalesforceMarketingCloudSource - from ._models_py3 import SalesforceObjectDataset - from ._models_py3 import SalesforceServiceCloudLinkedService - from ._models_py3 import SalesforceServiceCloudObjectDataset - from ._models_py3 import SalesforceServiceCloudSink - from ._models_py3 import SalesforceServiceCloudSource - from ._models_py3 import SalesforceSink - from ._models_py3 import SalesforceSource - from ._models_py3 import SapBWLinkedService - from ._models_py3 import SapBwCubeDataset - from ._models_py3 import SapBwSource - from ._models_py3 import SapCloudForCustomerLinkedService - from ._models_py3 import SapCloudForCustomerResourceDataset - from ._models_py3 import SapCloudForCustomerSink - from ._models_py3 import SapCloudForCustomerSource - from ._models_py3 import SapEccLinkedService - from ._models_py3 import SapEccResourceDataset - from ._models_py3 import SapEccSource - from ._models_py3 import SapHanaLinkedService - from ._models_py3 import SapHanaPartitionSettings - from ._models_py3 import SapHanaSource - from ._models_py3 import SapHanaTableDataset - from ._models_py3 import SapOpenHubLinkedService - from ._models_py3 import SapOpenHubSource - from ._models_py3 import SapOpenHubTableDataset - from ._models_py3 import SapTableLinkedService - from ._models_py3 import SapTablePartitionSettings - from ._models_py3 import SapTableResourceDataset - from ._models_py3 import SapTableSource - from ._models_py3 import ScheduleTrigger - from ._models_py3 import ScheduleTriggerRecurrence - from ._models_py3 import ScriptAction - from ._models_py3 import SecretBase - from ._models_py3 import SecureString - from ._models_py3 import SelfDependencyTumblingWindowTriggerReference - from ._models_py3 import SelfHostedIntegrationRuntime - from ._models_py3 import ServiceNowLinkedService - from ._models_py3 import ServiceNowObjectDataset - from ._models_py3 import ServiceNowSource - from ._models_py3 import SetVariableActivity - from ._models_py3 import SftpLocation - from ._models_py3 import SftpReadSettings - from ._models_py3 import SftpServerLinkedService - from ._models_py3 import SftpWriteSettings - from ._models_py3 import SharePointOnlineListLinkedService - from ._models_py3 import SharePointOnlineListResourceDataset - from ._models_py3 import SharePointOnlineListSource - from ._models_py3 import ShopifyLinkedService - from ._models_py3 import ShopifyObjectDataset - from ._models_py3 import ShopifySource - from ._models_py3 import SkipErrorFile - from ._models_py3 import Sku - from ._models_py3 import SnowflakeDataset - from ._models_py3 import SnowflakeExportCopyCommand - from ._models_py3 import SnowflakeImportCopyCommand - from ._models_py3 import SnowflakeLinkedService - from ._models_py3 import SnowflakeSink - from ._models_py3 import SnowflakeSource - from ._models_py3 import SparkBatchJob - from ._models_py3 import SparkBatchJobState - from ._models_py3 import SparkJobDefinition - from ._models_py3 import SparkJobDefinitionResource - from ._models_py3 import SparkJobDefinitionsListResponse - from ._models_py3 import SparkJobProperties - from ._models_py3 import SparkLinkedService - from ._models_py3 import SparkObjectDataset - from ._models_py3 import SparkRequest - from ._models_py3 import SparkScheduler - from ._models_py3 import SparkServiceError - from ._models_py3 import SparkServicePlugin - from ._models_py3 import SparkSource - from ._models_py3 import SqlConnection - from ._models_py3 import SqlDWSink - from ._models_py3 import SqlDWSource - from ._models_py3 import SqlMISink - from ._models_py3 import SqlMISource - from ._models_py3 import SqlPartitionSettings - from ._models_py3 import SqlPool - from ._models_py3 import SqlPoolInfoListResult - from ._models_py3 import SqlPoolReference - from ._models_py3 import SqlPoolStoredProcedureActivity - from ._models_py3 import SqlScript - from ._models_py3 import SqlScriptContent - from ._models_py3 import SqlScriptMetadata - from ._models_py3 import SqlScriptResource - from ._models_py3 import SqlScriptsListResponse - from ._models_py3 import SqlServerLinkedService - from ._models_py3 import SqlServerSink - from ._models_py3 import SqlServerSource - from ._models_py3 import SqlServerStoredProcedureActivity - from ._models_py3 import SqlServerTableDataset - from ._models_py3 import SqlSink - from ._models_py3 import SqlSource - from ._models_py3 import SquareLinkedService - from ._models_py3 import SquareObjectDataset - from ._models_py3 import SquareSource - from ._models_py3 import SsisObjectMetadataStatusResponse - from ._models_py3 import StagingSettings - from ._models_py3 import StartDataFlowDebugSessionRequest - from ._models_py3 import StartDataFlowDebugSessionResponse - from ._models_py3 import StoreReadSettings - from ._models_py3 import StoreWriteSettings - from ._models_py3 import StoredProcedureParameter - from ._models_py3 import SubResource - from ._models_py3 import SubResourceDebugResource - from ._models_py3 import SwitchActivity - from ._models_py3 import SwitchCase - from ._models_py3 import SybaseLinkedService - from ._models_py3 import SybaseSource - from ._models_py3 import SybaseTableDataset - from ._models_py3 import SynapseNotebookActivity - from ._models_py3 import SynapseNotebookReference - from ._models_py3 import SynapseSparkJobDefinitionActivity - from ._models_py3 import SynapseSparkJobReference - from ._models_py3 import TabularSource - from ._models_py3 import TabularTranslator - from ._models_py3 import TarGZipReadSettings - from ._models_py3 import TarReadSettings - from ._models_py3 import TeradataLinkedService - from ._models_py3 import TeradataPartitionSettings - from ._models_py3 import TeradataSource - from ._models_py3 import TeradataTableDataset - from ._models_py3 import TextFormat - from ._models_py3 import TrackedResource - from ._models_py3 import Transformation - from ._models_py3 import Trigger - from ._models_py3 import TriggerDependencyProvisioningStatus - from ._models_py3 import TriggerDependencyReference - from ._models_py3 import TriggerListResponse - from ._models_py3 import TriggerPipelineReference - from ._models_py3 import TriggerReference - from ._models_py3 import TriggerResource - from ._models_py3 import TriggerRun - from ._models_py3 import TriggerRunsQueryResponse - from ._models_py3 import TriggerSubscriptionOperationStatus - from ._models_py3 import TumblingWindowTrigger - from ._models_py3 import TumblingWindowTriggerDependencyReference - from ._models_py3 import TypeConversionSettings - from ._models_py3 import UntilActivity - from ._models_py3 import UserProperty - from ._models_py3 import ValidationActivity - from ._models_py3 import VariableSpecification - from ._models_py3 import VerticaLinkedService - from ._models_py3 import VerticaSource - from ._models_py3 import VerticaTableDataset - from ._models_py3 import VirtualNetworkProfile - from ._models_py3 import WaitActivity - from ._models_py3 import WebActivity - from ._models_py3 import WebActivityAuthentication - from ._models_py3 import WebAnonymousAuthentication - from ._models_py3 import WebBasicAuthentication - from ._models_py3 import WebClientCertificateAuthentication - from ._models_py3 import WebHookActivity - from ._models_py3 import WebLinkedService - from ._models_py3 import WebLinkedServiceTypeProperties - from ._models_py3 import WebSource - from ._models_py3 import WebTableDataset - from ._models_py3 import Workspace - from ._models_py3 import WorkspaceIdentity - from ._models_py3 import WorkspaceKeyDetails - from ._models_py3 import WorkspaceRepositoryConfiguration - from ._models_py3 import WorkspaceUpdateParameters - from ._models_py3 import XeroLinkedService - from ._models_py3 import XeroObjectDataset - from ._models_py3 import XeroSource - from ._models_py3 import XmlDataset - from ._models_py3 import XmlReadSettings - from ._models_py3 import XmlSource - from ._models_py3 import ZipDeflateReadSettings - from ._models_py3 import ZohoLinkedService - from ._models_py3 import ZohoObjectDataset - from ._models_py3 import ZohoSource -except (SyntaxError, ImportError): - from ._models import Activity # type: ignore - from ._models import ActivityDependency # type: ignore - from ._models import ActivityPolicy # type: ignore - from ._models import ActivityRun # type: ignore - from ._models import ActivityRunsQueryResponse # type: ignore - from ._models import AddDataFlowToDebugSessionResponse # type: ignore - from ._models import AdditionalColumns # type: ignore - from ._models import AmazonMWSLinkedService # type: ignore - from ._models import AmazonMWSObjectDataset # type: ignore - from ._models import AmazonMWSSource # type: ignore - from ._models import AmazonRedshiftLinkedService # type: ignore - from ._models import AmazonRedshiftSource # type: ignore - from ._models import AmazonRedshiftTableDataset # type: ignore - from ._models import AmazonS3Dataset # type: ignore - from ._models import AmazonS3LinkedService # type: ignore - from ._models import AmazonS3Location # type: ignore - from ._models import AmazonS3ReadSettings # type: ignore - from ._models import AppendVariableActivity # type: ignore - from ._models import ArtifactRenameRequest # type: ignore - from ._models import AutoPauseProperties # type: ignore - from ._models import AutoScaleProperties # type: ignore - from ._models import AvroDataset # type: ignore - from ._models import AvroFormat # type: ignore - from ._models import AvroSink # type: ignore - from ._models import AvroSource # type: ignore - from ._models import AvroWriteSettings # type: ignore - from ._models import AzureBatchLinkedService # type: ignore - from ._models import AzureBlobDataset # type: ignore - from ._models import AzureBlobFSDataset # type: ignore - from ._models import AzureBlobFSLinkedService # type: ignore - from ._models import AzureBlobFSLocation # type: ignore - from ._models import AzureBlobFSReadSettings # type: ignore - from ._models import AzureBlobFSSink # type: ignore - from ._models import AzureBlobFSSource # type: ignore - from ._models import AzureBlobFSWriteSettings # type: ignore - from ._models import AzureBlobStorageLinkedService # type: ignore - from ._models import AzureBlobStorageLocation # type: ignore - from ._models import AzureBlobStorageReadSettings # type: ignore - from ._models import AzureBlobStorageWriteSettings # type: ignore - from ._models import AzureDataExplorerCommandActivity # type: ignore - from ._models import AzureDataExplorerLinkedService # type: ignore - from ._models import AzureDataExplorerSink # type: ignore - from ._models import AzureDataExplorerSource # type: ignore - from ._models import AzureDataExplorerTableDataset # type: ignore - from ._models import AzureDataLakeAnalyticsLinkedService # type: ignore - from ._models import AzureDataLakeStoreDataset # type: ignore - from ._models import AzureDataLakeStoreLinkedService # type: ignore - from ._models import AzureDataLakeStoreLocation # type: ignore - from ._models import AzureDataLakeStoreReadSettings # type: ignore - from ._models import AzureDataLakeStoreSink # type: ignore - from ._models import AzureDataLakeStoreSource # type: ignore - from ._models import AzureDataLakeStoreWriteSettings # type: ignore - from ._models import AzureDatabricksDeltaLakeDataset # type: ignore - from ._models import AzureDatabricksDeltaLakeExportCommand # type: ignore - from ._models import AzureDatabricksDeltaLakeImportCommand # type: ignore - from ._models import AzureDatabricksDeltaLakeLinkedService # type: ignore - from ._models import AzureDatabricksDeltaLakeSink # type: ignore - from ._models import AzureDatabricksDeltaLakeSource # type: ignore - from ._models import AzureDatabricksLinkedService # type: ignore - from ._models import AzureEntityResource # type: ignore - from ._models import AzureFileStorageLinkedService # type: ignore - from ._models import AzureFileStorageLocation # type: ignore - from ._models import AzureFileStorageReadSettings # type: ignore - from ._models import AzureFileStorageWriteSettings # type: ignore - from ._models import AzureFunctionActivity # type: ignore - from ._models import AzureFunctionLinkedService # type: ignore - from ._models import AzureKeyVaultLinkedService # type: ignore - from ._models import AzureKeyVaultSecretReference # type: ignore - from ._models import AzureMLBatchExecutionActivity # type: ignore - from ._models import AzureMLExecutePipelineActivity # type: ignore - from ._models import AzureMLLinkedService # type: ignore - from ._models import AzureMLServiceLinkedService # type: ignore - from ._models import AzureMLUpdateResourceActivity # type: ignore - from ._models import AzureMLWebServiceFile # type: ignore - from ._models import AzureMariaDBLinkedService # type: ignore - from ._models import AzureMariaDBSource # type: ignore - from ._models import AzureMariaDBTableDataset # type: ignore - from ._models import AzureMySqlLinkedService # type: ignore - from ._models import AzureMySqlSink # type: ignore - from ._models import AzureMySqlSource # type: ignore - from ._models import AzureMySqlTableDataset # type: ignore - from ._models import AzurePostgreSqlLinkedService # type: ignore - from ._models import AzurePostgreSqlSink # type: ignore - from ._models import AzurePostgreSqlSource # type: ignore - from ._models import AzurePostgreSqlTableDataset # type: ignore - from ._models import AzureQueueSink # type: ignore - from ._models import AzureSearchIndexDataset # type: ignore - from ._models import AzureSearchIndexSink # type: ignore - from ._models import AzureSearchLinkedService # type: ignore - from ._models import AzureSqlDWLinkedService # type: ignore - from ._models import AzureSqlDWTableDataset # type: ignore - from ._models import AzureSqlDatabaseLinkedService # type: ignore - from ._models import AzureSqlMILinkedService # type: ignore - from ._models import AzureSqlMITableDataset # type: ignore - from ._models import AzureSqlSink # type: ignore - from ._models import AzureSqlSource # type: ignore - from ._models import AzureSqlTableDataset # type: ignore - from ._models import AzureStorageLinkedService # type: ignore - from ._models import AzureTableDataset # type: ignore - from ._models import AzureTableSink # type: ignore - from ._models import AzureTableSource # type: ignore - from ._models import AzureTableStorageLinkedService # type: ignore - from ._models import BigDataPoolReference # type: ignore - from ._models import BigDataPoolResourceInfo # type: ignore - from ._models import BigDataPoolResourceInfoListResult # type: ignore - from ._models import BinaryDataset # type: ignore - from ._models import BinaryReadSettings # type: ignore - from ._models import BinarySink # type: ignore - from ._models import BinarySource # type: ignore - from ._models import BlobEventsTrigger # type: ignore - from ._models import BlobSink # type: ignore - from ._models import BlobSource # type: ignore - from ._models import BlobTrigger # type: ignore - from ._models import CassandraLinkedService # type: ignore - from ._models import CassandraSource # type: ignore - from ._models import CassandraTableDataset # type: ignore - from ._models import ChainingTrigger # type: ignore - from ._models import CloudError # type: ignore - from ._models import CommonDataServiceForAppsEntityDataset # type: ignore - from ._models import CommonDataServiceForAppsLinkedService # type: ignore - from ._models import CommonDataServiceForAppsSink # type: ignore - from ._models import CommonDataServiceForAppsSource # type: ignore - from ._models import CompressionReadSettings # type: ignore - from ._models import ConcurLinkedService # type: ignore - from ._models import ConcurObjectDataset # type: ignore - from ._models import ConcurSource # type: ignore - from ._models import ControlActivity # type: ignore - from ._models import CopyActivity # type: ignore - from ._models import CopyActivityLogSettings # type: ignore - from ._models import CopySink # type: ignore - from ._models import CopySource # type: ignore - from ._models import CopyTranslator # type: ignore - from ._models import CosmosDbLinkedService # type: ignore - from ._models import CosmosDbMongoDbApiCollectionDataset # type: ignore - from ._models import CosmosDbMongoDbApiLinkedService # type: ignore - from ._models import CosmosDbMongoDbApiSink # type: ignore - from ._models import CosmosDbMongoDbApiSource # type: ignore - from ._models import CosmosDbSqlApiCollectionDataset # type: ignore - from ._models import CosmosDbSqlApiSink # type: ignore - from ._models import CosmosDbSqlApiSource # type: ignore - from ._models import CouchbaseLinkedService # type: ignore - from ._models import CouchbaseSource # type: ignore - from ._models import CouchbaseTableDataset # type: ignore - from ._models import CreateDataFlowDebugSessionRequest # type: ignore - from ._models import CreateDataFlowDebugSessionResponse # type: ignore - from ._models import CreateRunResponse # type: ignore - from ._models import CustomActivity # type: ignore - from ._models import CustomActivityReferenceObject # type: ignore - from ._models import CustomDataSourceLinkedService # type: ignore - from ._models import CustomDataset # type: ignore - from ._models import CustomEventsTrigger # type: ignore - from ._models import CustomSetupBase # type: ignore - from ._models import CustomerManagedKeyDetails # type: ignore - from ._models import DWCopyCommandDefaultValue # type: ignore - from ._models import DWCopyCommandSettings # type: ignore - from ._models import DataFlow # type: ignore - from ._models import DataFlowDebugCommandRequest # type: ignore - from ._models import DataFlowDebugCommandResponse # type: ignore - from ._models import DataFlowDebugPackage # type: ignore - from ._models import DataFlowDebugPackageDebugSettings # type: ignore - from ._models import DataFlowDebugPreviewDataRequest # type: ignore - from ._models import DataFlowDebugQueryResponse # type: ignore - from ._models import DataFlowDebugResource # type: ignore - from ._models import DataFlowDebugResultResponse # type: ignore - from ._models import DataFlowDebugSessionInfo # type: ignore - from ._models import DataFlowDebugStatisticsRequest # type: ignore - from ._models import DataFlowFolder # type: ignore - from ._models import DataFlowListResponse # type: ignore - from ._models import DataFlowReference # type: ignore - from ._models import DataFlowResource # type: ignore - from ._models import DataFlowSink # type: ignore - from ._models import DataFlowSource # type: ignore - from ._models import DataFlowSourceSetting # type: ignore - from ._models import DataFlowStagingInfo # type: ignore - from ._models import DataLakeAnalyticsUSQLActivity # type: ignore - from ._models import DataLakeStorageAccountDetails # type: ignore - from ._models import DatabricksNotebookActivity # type: ignore - from ._models import DatabricksSparkJarActivity # type: ignore - from ._models import DatabricksSparkPythonActivity # type: ignore - from ._models import Dataset # type: ignore - from ._models import DatasetBZip2Compression # type: ignore - from ._models import DatasetCompression # type: ignore - from ._models import DatasetDataElement # type: ignore - from ._models import DatasetDebugResource # type: ignore - from ._models import DatasetDeflateCompression # type: ignore - from ._models import DatasetFolder # type: ignore - from ._models import DatasetGZipCompression # type: ignore - from ._models import DatasetListResponse # type: ignore - from ._models import DatasetLocation # type: ignore - from ._models import DatasetReference # type: ignore - from ._models import DatasetResource # type: ignore - from ._models import DatasetSchemaDataElement # type: ignore - from ._models import DatasetStorageFormat # type: ignore - from ._models import DatasetTarCompression # type: ignore - from ._models import DatasetTarGZipCompression # type: ignore - from ._models import DatasetZipDeflateCompression # type: ignore - from ._models import Db2LinkedService # type: ignore - from ._models import Db2Source # type: ignore - from ._models import Db2TableDataset # type: ignore - from ._models import DeleteActivity # type: ignore - from ._models import DeleteDataFlowDebugSessionRequest # type: ignore - from ._models import DelimitedTextDataset # type: ignore - from ._models import DelimitedTextReadSettings # type: ignore - from ._models import DelimitedTextSink # type: ignore - from ._models import DelimitedTextSource # type: ignore - from ._models import DelimitedTextWriteSettings # type: ignore - from ._models import DependencyReference # type: ignore - from ._models import DistcpSettings # type: ignore - from ._models import DocumentDbCollectionDataset # type: ignore - from ._models import DocumentDbCollectionSink # type: ignore - from ._models import DocumentDbCollectionSource # type: ignore - from ._models import DrillLinkedService # type: ignore - from ._models import DrillSource # type: ignore - from ._models import DrillTableDataset # type: ignore - from ._models import DynamicExecutorAllocation # type: ignore - from ._models import DynamicsAXLinkedService # type: ignore - from ._models import DynamicsAXResourceDataset # type: ignore - from ._models import DynamicsAXSource # type: ignore - from ._models import DynamicsCrmEntityDataset # type: ignore - from ._models import DynamicsCrmLinkedService # type: ignore - from ._models import DynamicsCrmSink # type: ignore - from ._models import DynamicsCrmSource # type: ignore - from ._models import DynamicsEntityDataset # type: ignore - from ._models import DynamicsLinkedService # type: ignore - from ._models import DynamicsSink # type: ignore - from ._models import DynamicsSource # type: ignore - from ._models import EloquaLinkedService # type: ignore - from ._models import EloquaObjectDataset # type: ignore - from ._models import EloquaSource # type: ignore - from ._models import EncryptionDetails # type: ignore - from ._models import EntityReference # type: ignore - from ._models import ErrorAdditionalInfo # type: ignore - from ._models import ErrorContract # type: ignore - from ._models import ErrorResponse # type: ignore - from ._models import EvaluateDataFlowExpressionRequest # type: ignore - from ._models import ExcelDataset # type: ignore - from ._models import ExcelSource # type: ignore - from ._models import ExecuteDataFlowActivity # type: ignore - from ._models import ExecuteDataFlowActivityTypePropertiesCompute # type: ignore - from ._models import ExecutePipelineActivity # type: ignore - from ._models import ExecuteSSISPackageActivity # type: ignore - from ._models import ExecutionActivity # type: ignore - from ._models import ExportSettings # type: ignore - from ._models import ExposureControlRequest # type: ignore - from ._models import ExposureControlResponse # type: ignore - from ._models import Expression # type: ignore - from ._models import FileServerLinkedService # type: ignore - from ._models import FileServerLocation # type: ignore - from ._models import FileServerReadSettings # type: ignore - from ._models import FileServerWriteSettings # type: ignore - from ._models import FileShareDataset # type: ignore - from ._models import FileSystemSink # type: ignore - from ._models import FileSystemSource # type: ignore - from ._models import FilterActivity # type: ignore - from ._models import ForEachActivity # type: ignore - from ._models import FormatReadSettings # type: ignore - from ._models import FormatWriteSettings # type: ignore - from ._models import FtpReadSettings # type: ignore - from ._models import FtpServerLinkedService # type: ignore - from ._models import FtpServerLocation # type: ignore - from ._models import GetMetadataActivity # type: ignore - from ._models import GetSsisObjectMetadataRequest # type: ignore - from ._models import GitHubAccessTokenRequest # type: ignore - from ._models import GitHubAccessTokenResponse # type: ignore - from ._models import GitHubClientSecret # type: ignore - from ._models import GoogleAdWordsLinkedService # type: ignore - from ._models import GoogleAdWordsObjectDataset # type: ignore - from ._models import GoogleAdWordsSource # type: ignore - from ._models import GoogleBigQueryLinkedService # type: ignore - from ._models import GoogleBigQueryObjectDataset # type: ignore - from ._models import GoogleBigQuerySource # type: ignore - from ._models import GoogleCloudStorageLinkedService # type: ignore - from ._models import GoogleCloudStorageLocation # type: ignore - from ._models import GoogleCloudStorageReadSettings # type: ignore - from ._models import GreenplumLinkedService # type: ignore - from ._models import GreenplumSource # type: ignore - from ._models import GreenplumTableDataset # type: ignore - from ._models import HBaseLinkedService # type: ignore - from ._models import HBaseObjectDataset # type: ignore - from ._models import HBaseSource # type: ignore - from ._models import HDInsightHiveActivity # type: ignore - from ._models import HDInsightLinkedService # type: ignore - from ._models import HDInsightMapReduceActivity # type: ignore - from ._models import HDInsightOnDemandLinkedService # type: ignore - from ._models import HDInsightPigActivity # type: ignore - from ._models import HDInsightSparkActivity # type: ignore - from ._models import HDInsightStreamingActivity # type: ignore - from ._models import HdfsLinkedService # type: ignore - from ._models import HdfsLocation # type: ignore - from ._models import HdfsReadSettings # type: ignore - from ._models import HdfsSource # type: ignore - from ._models import HiveLinkedService # type: ignore - from ._models import HiveObjectDataset # type: ignore - from ._models import HiveSource # type: ignore - from ._models import HttpDataset # type: ignore - from ._models import HttpLinkedService # type: ignore - from ._models import HttpReadSettings # type: ignore - from ._models import HttpServerLocation # type: ignore - from ._models import HttpSource # type: ignore - from ._models import HubspotLinkedService # type: ignore - from ._models import HubspotObjectDataset # type: ignore - from ._models import HubspotSource # type: ignore - from ._models import IfConditionActivity # type: ignore - from ._models import ImpalaLinkedService # type: ignore - from ._models import ImpalaObjectDataset # type: ignore - from ._models import ImpalaSource # type: ignore - from ._models import ImportSettings # type: ignore - from ._models import InformixLinkedService # type: ignore - from ._models import InformixSink # type: ignore - from ._models import InformixSource # type: ignore - from ._models import InformixTableDataset # type: ignore - from ._models import IntegrationRuntime # type: ignore - from ._models import IntegrationRuntimeComputeProperties # type: ignore - from ._models import IntegrationRuntimeCustomSetupScriptProperties # type: ignore - from ._models import IntegrationRuntimeDataFlowProperties # type: ignore - from ._models import IntegrationRuntimeDataProxyProperties # type: ignore - from ._models import IntegrationRuntimeListResponse # type: ignore - from ._models import IntegrationRuntimeReference # type: ignore - from ._models import IntegrationRuntimeResource # type: ignore - from ._models import IntegrationRuntimeSsisCatalogInfo # type: ignore - from ._models import IntegrationRuntimeSsisProperties # type: ignore - from ._models import IntegrationRuntimeVNetProperties # type: ignore - from ._models import JiraLinkedService # type: ignore - from ._models import JiraObjectDataset # type: ignore - from ._models import JiraSource # type: ignore - from ._models import JsonDataset # type: ignore - from ._models import JsonFormat # type: ignore - from ._models import JsonReadSettings # type: ignore - from ._models import JsonSink # type: ignore - from ._models import JsonSource # type: ignore - from ._models import JsonWriteSettings # type: ignore - from ._models import LibraryInfo # type: ignore - from ._models import LibraryListResponse # type: ignore - from ._models import LibraryRequirements # type: ignore - from ._models import LibraryResource # type: ignore - from ._models import LibraryResourceInfo # type: ignore - from ._models import LibraryResourceProperties # type: ignore - from ._models import LinkedIntegrationRuntimeKeyAuthorization # type: ignore - from ._models import LinkedIntegrationRuntimeRbacAuthorization # type: ignore - from ._models import LinkedIntegrationRuntimeType # type: ignore - from ._models import LinkedService # type: ignore - from ._models import LinkedServiceDebugResource # type: ignore - from ._models import LinkedServiceListResponse # type: ignore - from ._models import LinkedServiceReference # type: ignore - from ._models import LinkedServiceResource # type: ignore - from ._models import LogLocationSettings # type: ignore - from ._models import LogSettings # type: ignore - from ._models import LogStorageSettings # type: ignore - from ._models import LookupActivity # type: ignore - from ._models import MagentoLinkedService # type: ignore - from ._models import MagentoObjectDataset # type: ignore - from ._models import MagentoSource # type: ignore - from ._models import ManagedIdentity # type: ignore - from ._models import ManagedIntegrationRuntime # type: ignore - from ._models import ManagedVirtualNetworkReference # type: ignore - from ._models import ManagedVirtualNetworkSettings # type: ignore - from ._models import MappingDataFlow # type: ignore - from ._models import MariaDBLinkedService # type: ignore - from ._models import MariaDBSource # type: ignore - from ._models import MariaDBTableDataset # type: ignore - from ._models import MarketoLinkedService # type: ignore - from ._models import MarketoObjectDataset # type: ignore - from ._models import MarketoSource # type: ignore - from ._models import MicrosoftAccessLinkedService # type: ignore - from ._models import MicrosoftAccessSink # type: ignore - from ._models import MicrosoftAccessSource # type: ignore - from ._models import MicrosoftAccessTableDataset # type: ignore - from ._models import MongoDbAtlasCollectionDataset # type: ignore - from ._models import MongoDbAtlasLinkedService # type: ignore - from ._models import MongoDbAtlasSource # type: ignore - from ._models import MongoDbCollectionDataset # type: ignore - from ._models import MongoDbCursorMethodsProperties # type: ignore - from ._models import MongoDbLinkedService # type: ignore - from ._models import MongoDbSource # type: ignore - from ._models import MongoDbV2CollectionDataset # type: ignore - from ._models import MongoDbV2LinkedService # type: ignore - from ._models import MongoDbV2Source # type: ignore - from ._models import MultiplePipelineTrigger # type: ignore - from ._models import MySqlLinkedService # type: ignore - from ._models import MySqlSource # type: ignore - from ._models import MySqlTableDataset # type: ignore - from ._models import NetezzaLinkedService # type: ignore - from ._models import NetezzaPartitionSettings # type: ignore - from ._models import NetezzaSource # type: ignore - from ._models import NetezzaTableDataset # type: ignore - from ._models import Notebook # type: ignore - from ._models import NotebookCell # type: ignore - from ._models import NotebookCellOutputItem # type: ignore - from ._models import NotebookKernelSpec # type: ignore - from ._models import NotebookLanguageInfo # type: ignore - from ._models import NotebookListResponse # type: ignore - from ._models import NotebookMetadata # type: ignore - from ._models import NotebookResource # type: ignore - from ._models import NotebookSessionProperties # type: ignore - from ._models import ODataLinkedService # type: ignore - from ._models import ODataResourceDataset # type: ignore - from ._models import ODataSource # type: ignore - from ._models import OdbcLinkedService # type: ignore - from ._models import OdbcSink # type: ignore - from ._models import OdbcSource # type: ignore - from ._models import OdbcTableDataset # type: ignore - from ._models import Office365Dataset # type: ignore - from ._models import Office365LinkedService # type: ignore - from ._models import Office365Source # type: ignore - from ._models import OperationResult # type: ignore - from ._models import OracleLinkedService # type: ignore - from ._models import OraclePartitionSettings # type: ignore - from ._models import OracleServiceCloudLinkedService # type: ignore - from ._models import OracleServiceCloudObjectDataset # type: ignore - from ._models import OracleServiceCloudSource # type: ignore - from ._models import OracleSink # type: ignore - from ._models import OracleSource # type: ignore - from ._models import OracleTableDataset # type: ignore - from ._models import OrcDataset # type: ignore - from ._models import OrcFormat # type: ignore - from ._models import OrcSink # type: ignore - from ._models import OrcSource # type: ignore - from ._models import OrcWriteSettings # type: ignore - from ._models import ParameterSpecification # type: ignore - from ._models import ParquetDataset # type: ignore - from ._models import ParquetFormat # type: ignore - from ._models import ParquetSink # type: ignore - from ._models import ParquetSource # type: ignore - from ._models import ParquetWriteSettings # type: ignore - from ._models import PaypalLinkedService # type: ignore - from ._models import PaypalObjectDataset # type: ignore - from ._models import PaypalSource # type: ignore - from ._models import PhoenixLinkedService # type: ignore - from ._models import PhoenixObjectDataset # type: ignore - from ._models import PhoenixSource # type: ignore - from ._models import PipelineFolder # type: ignore - from ._models import PipelineListResponse # type: ignore - from ._models import PipelineReference # type: ignore - from ._models import PipelineResource # type: ignore - from ._models import PipelineRun # type: ignore - from ._models import PipelineRunInvokedBy # type: ignore - from ._models import PipelineRunsQueryResponse # type: ignore - from ._models import PolybaseSettings # type: ignore - from ._models import PostgreSqlLinkedService # type: ignore - from ._models import PostgreSqlSource # type: ignore - from ._models import PostgreSqlTableDataset # type: ignore - from ._models import PrestoLinkedService # type: ignore - from ._models import PrestoObjectDataset # type: ignore - from ._models import PrestoSource # type: ignore - from ._models import PrivateEndpoint # type: ignore - from ._models import PrivateEndpointConnection # type: ignore - from ._models import PrivateLinkServiceConnectionState # type: ignore - from ._models import ProxyResource # type: ignore - from ._models import PurviewConfiguration # type: ignore - from ._models import QueryDataFlowDebugSessionsResponse # type: ignore - from ._models import QuickBooksLinkedService # type: ignore - from ._models import QuickBooksObjectDataset # type: ignore - from ._models import QuickBooksSource # type: ignore - from ._models import RecurrenceSchedule # type: ignore - from ._models import RecurrenceScheduleOccurrence # type: ignore - from ._models import RedirectIncompatibleRowSettings # type: ignore - from ._models import RedshiftUnloadSettings # type: ignore - from ._models import RelationalSource # type: ignore - from ._models import RelationalTableDataset # type: ignore - from ._models import RerunTriggerListResponse # type: ignore - from ._models import RerunTriggerResource # type: ignore - from ._models import RerunTumblingWindowTrigger # type: ignore - from ._models import RerunTumblingWindowTriggerActionParameters # type: ignore - from ._models import Resource # type: ignore - from ._models import ResponsysLinkedService # type: ignore - from ._models import ResponsysObjectDataset # type: ignore - from ._models import ResponsysSource # type: ignore - from ._models import RestResourceDataset # type: ignore - from ._models import RestServiceLinkedService # type: ignore - from ._models import RestSink # type: ignore - from ._models import RestSource # type: ignore - from ._models import RetryPolicy # type: ignore - from ._models import RunFilterParameters # type: ignore - from ._models import RunQueryFilter # type: ignore - from ._models import RunQueryOrderBy # type: ignore - from ._models import SSISAccessCredential # type: ignore - from ._models import SSISChildPackage # type: ignore - from ._models import SSISExecutionCredential # type: ignore - from ._models import SSISExecutionParameter # type: ignore - from ._models import SSISLogLocation # type: ignore - from ._models import SSISPackageLocation # type: ignore - from ._models import SSISPropertyOverride # type: ignore - from ._models import SalesforceLinkedService # type: ignore - from ._models import SalesforceMarketingCloudLinkedService # type: ignore - from ._models import SalesforceMarketingCloudObjectDataset # type: ignore - from ._models import SalesforceMarketingCloudSource # type: ignore - from ._models import SalesforceObjectDataset # type: ignore - from ._models import SalesforceServiceCloudLinkedService # type: ignore - from ._models import SalesforceServiceCloudObjectDataset # type: ignore - from ._models import SalesforceServiceCloudSink # type: ignore - from ._models import SalesforceServiceCloudSource # type: ignore - from ._models import SalesforceSink # type: ignore - from ._models import SalesforceSource # type: ignore - from ._models import SapBWLinkedService # type: ignore - from ._models import SapBwCubeDataset # type: ignore - from ._models import SapBwSource # type: ignore - from ._models import SapCloudForCustomerLinkedService # type: ignore - from ._models import SapCloudForCustomerResourceDataset # type: ignore - from ._models import SapCloudForCustomerSink # type: ignore - from ._models import SapCloudForCustomerSource # type: ignore - from ._models import SapEccLinkedService # type: ignore - from ._models import SapEccResourceDataset # type: ignore - from ._models import SapEccSource # type: ignore - from ._models import SapHanaLinkedService # type: ignore - from ._models import SapHanaPartitionSettings # type: ignore - from ._models import SapHanaSource # type: ignore - from ._models import SapHanaTableDataset # type: ignore - from ._models import SapOpenHubLinkedService # type: ignore - from ._models import SapOpenHubSource # type: ignore - from ._models import SapOpenHubTableDataset # type: ignore - from ._models import SapTableLinkedService # type: ignore - from ._models import SapTablePartitionSettings # type: ignore - from ._models import SapTableResourceDataset # type: ignore - from ._models import SapTableSource # type: ignore - from ._models import ScheduleTrigger # type: ignore - from ._models import ScheduleTriggerRecurrence # type: ignore - from ._models import ScriptAction # type: ignore - from ._models import SecretBase # type: ignore - from ._models import SecureString # type: ignore - from ._models import SelfDependencyTumblingWindowTriggerReference # type: ignore - from ._models import SelfHostedIntegrationRuntime # type: ignore - from ._models import ServiceNowLinkedService # type: ignore - from ._models import ServiceNowObjectDataset # type: ignore - from ._models import ServiceNowSource # type: ignore - from ._models import SetVariableActivity # type: ignore - from ._models import SftpLocation # type: ignore - from ._models import SftpReadSettings # type: ignore - from ._models import SftpServerLinkedService # type: ignore - from ._models import SftpWriteSettings # type: ignore - from ._models import SharePointOnlineListLinkedService # type: ignore - from ._models import SharePointOnlineListResourceDataset # type: ignore - from ._models import SharePointOnlineListSource # type: ignore - from ._models import ShopifyLinkedService # type: ignore - from ._models import ShopifyObjectDataset # type: ignore - from ._models import ShopifySource # type: ignore - from ._models import SkipErrorFile # type: ignore - from ._models import Sku # type: ignore - from ._models import SnowflakeDataset # type: ignore - from ._models import SnowflakeExportCopyCommand # type: ignore - from ._models import SnowflakeImportCopyCommand # type: ignore - from ._models import SnowflakeLinkedService # type: ignore - from ._models import SnowflakeSink # type: ignore - from ._models import SnowflakeSource # type: ignore - from ._models import SparkBatchJob # type: ignore - from ._models import SparkBatchJobState # type: ignore - from ._models import SparkJobDefinition # type: ignore - from ._models import SparkJobDefinitionResource # type: ignore - from ._models import SparkJobDefinitionsListResponse # type: ignore - from ._models import SparkJobProperties # type: ignore - from ._models import SparkLinkedService # type: ignore - from ._models import SparkObjectDataset # type: ignore - from ._models import SparkRequest # type: ignore - from ._models import SparkScheduler # type: ignore - from ._models import SparkServiceError # type: ignore - from ._models import SparkServicePlugin # type: ignore - from ._models import SparkSource # type: ignore - from ._models import SqlConnection # type: ignore - from ._models import SqlDWSink # type: ignore - from ._models import SqlDWSource # type: ignore - from ._models import SqlMISink # type: ignore - from ._models import SqlMISource # type: ignore - from ._models import SqlPartitionSettings # type: ignore - from ._models import SqlPool # type: ignore - from ._models import SqlPoolInfoListResult # type: ignore - from ._models import SqlPoolReference # type: ignore - from ._models import SqlPoolStoredProcedureActivity # type: ignore - from ._models import SqlScript # type: ignore - from ._models import SqlScriptContent # type: ignore - from ._models import SqlScriptMetadata # type: ignore - from ._models import SqlScriptResource # type: ignore - from ._models import SqlScriptsListResponse # type: ignore - from ._models import SqlServerLinkedService # type: ignore - from ._models import SqlServerSink # type: ignore - from ._models import SqlServerSource # type: ignore - from ._models import SqlServerStoredProcedureActivity # type: ignore - from ._models import SqlServerTableDataset # type: ignore - from ._models import SqlSink # type: ignore - from ._models import SqlSource # type: ignore - from ._models import SquareLinkedService # type: ignore - from ._models import SquareObjectDataset # type: ignore - from ._models import SquareSource # type: ignore - from ._models import SsisObjectMetadataStatusResponse # type: ignore - from ._models import StagingSettings # type: ignore - from ._models import StartDataFlowDebugSessionRequest # type: ignore - from ._models import StartDataFlowDebugSessionResponse # type: ignore - from ._models import StoreReadSettings # type: ignore - from ._models import StoreWriteSettings # type: ignore - from ._models import StoredProcedureParameter # type: ignore - from ._models import SubResource # type: ignore - from ._models import SubResourceDebugResource # type: ignore - from ._models import SwitchActivity # type: ignore - from ._models import SwitchCase # type: ignore - from ._models import SybaseLinkedService # type: ignore - from ._models import SybaseSource # type: ignore - from ._models import SybaseTableDataset # type: ignore - from ._models import SynapseNotebookActivity # type: ignore - from ._models import SynapseNotebookReference # type: ignore - from ._models import SynapseSparkJobDefinitionActivity # type: ignore - from ._models import SynapseSparkJobReference # type: ignore - from ._models import TabularSource # type: ignore - from ._models import TabularTranslator # type: ignore - from ._models import TarGZipReadSettings # type: ignore - from ._models import TarReadSettings # type: ignore - from ._models import TeradataLinkedService # type: ignore - from ._models import TeradataPartitionSettings # type: ignore - from ._models import TeradataSource # type: ignore - from ._models import TeradataTableDataset # type: ignore - from ._models import TextFormat # type: ignore - from ._models import TrackedResource # type: ignore - from ._models import Transformation # type: ignore - from ._models import Trigger # type: ignore - from ._models import TriggerDependencyProvisioningStatus # type: ignore - from ._models import TriggerDependencyReference # type: ignore - from ._models import TriggerListResponse # type: ignore - from ._models import TriggerPipelineReference # type: ignore - from ._models import TriggerReference # type: ignore - from ._models import TriggerResource # type: ignore - from ._models import TriggerRun # type: ignore - from ._models import TriggerRunsQueryResponse # type: ignore - from ._models import TriggerSubscriptionOperationStatus # type: ignore - from ._models import TumblingWindowTrigger # type: ignore - from ._models import TumblingWindowTriggerDependencyReference # type: ignore - from ._models import TypeConversionSettings # type: ignore - from ._models import UntilActivity # type: ignore - from ._models import UserProperty # type: ignore - from ._models import ValidationActivity # type: ignore - from ._models import VariableSpecification # type: ignore - from ._models import VerticaLinkedService # type: ignore - from ._models import VerticaSource # type: ignore - from ._models import VerticaTableDataset # type: ignore - from ._models import VirtualNetworkProfile # type: ignore - from ._models import WaitActivity # type: ignore - from ._models import WebActivity # type: ignore - from ._models import WebActivityAuthentication # type: ignore - from ._models import WebAnonymousAuthentication # type: ignore - from ._models import WebBasicAuthentication # type: ignore - from ._models import WebClientCertificateAuthentication # type: ignore - from ._models import WebHookActivity # type: ignore - from ._models import WebLinkedService # type: ignore - from ._models import WebLinkedServiceTypeProperties # type: ignore - from ._models import WebSource # type: ignore - from ._models import WebTableDataset # type: ignore - from ._models import Workspace # type: ignore - from ._models import WorkspaceIdentity # type: ignore - from ._models import WorkspaceKeyDetails # type: ignore - from ._models import WorkspaceRepositoryConfiguration # type: ignore - from ._models import WorkspaceUpdateParameters # type: ignore - from ._models import XeroLinkedService # type: ignore - from ._models import XeroObjectDataset # type: ignore - from ._models import XeroSource # type: ignore - from ._models import XmlDataset # type: ignore - from ._models import XmlReadSettings # type: ignore - from ._models import XmlSource # type: ignore - from ._models import ZipDeflateReadSettings # type: ignore - from ._models import ZohoLinkedService # type: ignore - from ._models import ZohoObjectDataset # type: ignore - from ._models import ZohoSource # type: ignore - -from ._artifacts_client_enums import ( - AvroCompressionCodec, - AzureFunctionActivityMethod, - AzureSearchIndexWriteBehaviorType, - BigDataPoolReferenceType, - BlobEventType, - CassandraSourceReadConsistencyLevels, - CellOutputType, - CompressionCodec, - CopyBehaviorType, - DataFlowComputeType, - DataFlowReferenceType, - DatasetCompressionLevel, - DatasetReferenceType, - DayOfWeek, - Db2AuthenticationType, - DependencyCondition, - DynamicsAuthenticationType, - DynamicsDeploymentType, - DynamicsServicePrincipalCredentialType, - DynamicsSinkWriteBehavior, - EventSubscriptionStatus, - ExpressionType, - FtpAuthenticationType, - GoogleAdWordsAuthenticationType, - GoogleBigQueryAuthenticationType, - HBaseAuthenticationType, - HDInsightActivityDebugInfoOption, - HdiNodeTypes, - HiveAuthenticationType, - HiveServerType, - HiveThriftTransportProtocol, - HttpAuthenticationType, - ImpalaAuthenticationType, - IntegrationRuntimeEdition, - IntegrationRuntimeEntityReferenceType, - IntegrationRuntimeLicenseType, - IntegrationRuntimeReferenceType, - IntegrationRuntimeSsisCatalogPricingTier, - IntegrationRuntimeState, - IntegrationRuntimeType, - JsonFormatFilePattern, - JsonWriteFilePattern, - MongoDbAuthenticationType, - NetezzaPartitionOption, - NodeSize, - NodeSizeFamily, - NotebookReferenceType, - ODataAadServicePrincipalCredentialType, - ODataAuthenticationType, - OraclePartitionOption, - OrcCompressionCodec, - ParameterType, - ParquetCompressionCodecEnum, - PhoenixAuthenticationType, - PipelineReferenceType, - PluginCurrentState, - PolybaseSettingsRejectType, - PrestoAuthenticationType, - RecurrenceFrequency, - ResourceIdentityType, - RestServiceAuthenticationType, - RunQueryFilterOperand, - RunQueryFilterOperator, - RunQueryOrder, - RunQueryOrderByField, - SalesforceSinkWriteBehavior, - SalesforceSourceReadBehavior, - SapCloudForCustomerSinkWriteBehavior, - SapHanaAuthenticationType, - SapHanaPartitionOption, - SapTablePartitionOption, - SchedulerCurrentState, - ServiceNowAuthenticationType, - SftpAuthenticationType, - SparkAuthenticationType, - SparkBatchJobResultType, - SparkErrorSource, - SparkJobReferenceType, - SparkJobType, - SparkServerType, - SparkThriftTransportProtocol, - SqlConnectionType, - SqlPartitionOption, - SqlPoolReferenceType, - SqlScriptType, - SsisLogLocationType, - SsisPackageLocationType, - StoredProcedureParameterType, - SybaseAuthenticationType, - TeradataAuthenticationType, - TeradataPartitionOption, - TriggerReferenceType, - TriggerRunStatus, - TriggerRuntimeState, - TumblingWindowFrequency, - Type, - VariableType, - WebActivityMethod, - WebAuthenticationType, - WebHookActivityMethod, -) - -__all__ = [ - 'Activity', - 'ActivityDependency', - 'ActivityPolicy', - 'ActivityRun', - 'ActivityRunsQueryResponse', - 'AddDataFlowToDebugSessionResponse', - 'AdditionalColumns', - 'AmazonMWSLinkedService', - 'AmazonMWSObjectDataset', - 'AmazonMWSSource', - 'AmazonRedshiftLinkedService', - 'AmazonRedshiftSource', - 'AmazonRedshiftTableDataset', - 'AmazonS3Dataset', - 'AmazonS3LinkedService', - 'AmazonS3Location', - 'AmazonS3ReadSettings', - 'AppendVariableActivity', - 'ArtifactRenameRequest', - 'AutoPauseProperties', - 'AutoScaleProperties', - 'AvroDataset', - 'AvroFormat', - 'AvroSink', - 'AvroSource', - 'AvroWriteSettings', - 'AzureBatchLinkedService', - 'AzureBlobDataset', - 'AzureBlobFSDataset', - 'AzureBlobFSLinkedService', - 'AzureBlobFSLocation', - 'AzureBlobFSReadSettings', - 'AzureBlobFSSink', - 'AzureBlobFSSource', - 'AzureBlobFSWriteSettings', - 'AzureBlobStorageLinkedService', - 'AzureBlobStorageLocation', - 'AzureBlobStorageReadSettings', - 'AzureBlobStorageWriteSettings', - 'AzureDataExplorerCommandActivity', - 'AzureDataExplorerLinkedService', - 'AzureDataExplorerSink', - 'AzureDataExplorerSource', - 'AzureDataExplorerTableDataset', - 'AzureDataLakeAnalyticsLinkedService', - 'AzureDataLakeStoreDataset', - 'AzureDataLakeStoreLinkedService', - 'AzureDataLakeStoreLocation', - 'AzureDataLakeStoreReadSettings', - 'AzureDataLakeStoreSink', - 'AzureDataLakeStoreSource', - 'AzureDataLakeStoreWriteSettings', - 'AzureDatabricksDeltaLakeDataset', - 'AzureDatabricksDeltaLakeExportCommand', - 'AzureDatabricksDeltaLakeImportCommand', - 'AzureDatabricksDeltaLakeLinkedService', - 'AzureDatabricksDeltaLakeSink', - 'AzureDatabricksDeltaLakeSource', - 'AzureDatabricksLinkedService', - 'AzureEntityResource', - 'AzureFileStorageLinkedService', - 'AzureFileStorageLocation', - 'AzureFileStorageReadSettings', - 'AzureFileStorageWriteSettings', - 'AzureFunctionActivity', - 'AzureFunctionLinkedService', - 'AzureKeyVaultLinkedService', - 'AzureKeyVaultSecretReference', - 'AzureMLBatchExecutionActivity', - 'AzureMLExecutePipelineActivity', - 'AzureMLLinkedService', - 'AzureMLServiceLinkedService', - 'AzureMLUpdateResourceActivity', - 'AzureMLWebServiceFile', - 'AzureMariaDBLinkedService', - 'AzureMariaDBSource', - 'AzureMariaDBTableDataset', - 'AzureMySqlLinkedService', - 'AzureMySqlSink', - 'AzureMySqlSource', - 'AzureMySqlTableDataset', - 'AzurePostgreSqlLinkedService', - 'AzurePostgreSqlSink', - 'AzurePostgreSqlSource', - 'AzurePostgreSqlTableDataset', - 'AzureQueueSink', - 'AzureSearchIndexDataset', - 'AzureSearchIndexSink', - 'AzureSearchLinkedService', - 'AzureSqlDWLinkedService', - 'AzureSqlDWTableDataset', - 'AzureSqlDatabaseLinkedService', - 'AzureSqlMILinkedService', - 'AzureSqlMITableDataset', - 'AzureSqlSink', - 'AzureSqlSource', - 'AzureSqlTableDataset', - 'AzureStorageLinkedService', - 'AzureTableDataset', - 'AzureTableSink', - 'AzureTableSource', - 'AzureTableStorageLinkedService', - 'BigDataPoolReference', - 'BigDataPoolResourceInfo', - 'BigDataPoolResourceInfoListResult', - 'BinaryDataset', - 'BinaryReadSettings', - 'BinarySink', - 'BinarySource', - 'BlobEventsTrigger', - 'BlobSink', - 'BlobSource', - 'BlobTrigger', - 'CassandraLinkedService', - 'CassandraSource', - 'CassandraTableDataset', - 'ChainingTrigger', - 'CloudError', - 'CommonDataServiceForAppsEntityDataset', - 'CommonDataServiceForAppsLinkedService', - 'CommonDataServiceForAppsSink', - 'CommonDataServiceForAppsSource', - 'CompressionReadSettings', - 'ConcurLinkedService', - 'ConcurObjectDataset', - 'ConcurSource', - 'ControlActivity', - 'CopyActivity', - 'CopyActivityLogSettings', - 'CopySink', - 'CopySource', - 'CopyTranslator', - 'CosmosDbLinkedService', - 'CosmosDbMongoDbApiCollectionDataset', - 'CosmosDbMongoDbApiLinkedService', - 'CosmosDbMongoDbApiSink', - 'CosmosDbMongoDbApiSource', - 'CosmosDbSqlApiCollectionDataset', - 'CosmosDbSqlApiSink', - 'CosmosDbSqlApiSource', - 'CouchbaseLinkedService', - 'CouchbaseSource', - 'CouchbaseTableDataset', - 'CreateDataFlowDebugSessionRequest', - 'CreateDataFlowDebugSessionResponse', - 'CreateRunResponse', - 'CustomActivity', - 'CustomActivityReferenceObject', - 'CustomDataSourceLinkedService', - 'CustomDataset', - 'CustomEventsTrigger', - 'CustomSetupBase', - 'CustomerManagedKeyDetails', - 'DWCopyCommandDefaultValue', - 'DWCopyCommandSettings', - 'DataFlow', - 'DataFlowDebugCommandRequest', - 'DataFlowDebugCommandResponse', - 'DataFlowDebugPackage', - 'DataFlowDebugPackageDebugSettings', - 'DataFlowDebugPreviewDataRequest', - 'DataFlowDebugQueryResponse', - 'DataFlowDebugResource', - 'DataFlowDebugResultResponse', - 'DataFlowDebugSessionInfo', - 'DataFlowDebugStatisticsRequest', - 'DataFlowFolder', - 'DataFlowListResponse', - 'DataFlowReference', - 'DataFlowResource', - 'DataFlowSink', - 'DataFlowSource', - 'DataFlowSourceSetting', - 'DataFlowStagingInfo', - 'DataLakeAnalyticsUSQLActivity', - 'DataLakeStorageAccountDetails', - 'DatabricksNotebookActivity', - 'DatabricksSparkJarActivity', - 'DatabricksSparkPythonActivity', - 'Dataset', - 'DatasetBZip2Compression', - 'DatasetCompression', - 'DatasetDataElement', - 'DatasetDebugResource', - 'DatasetDeflateCompression', - 'DatasetFolder', - 'DatasetGZipCompression', - 'DatasetListResponse', - 'DatasetLocation', - 'DatasetReference', - 'DatasetResource', - 'DatasetSchemaDataElement', - 'DatasetStorageFormat', - 'DatasetTarCompression', - 'DatasetTarGZipCompression', - 'DatasetZipDeflateCompression', - 'Db2LinkedService', - 'Db2Source', - 'Db2TableDataset', - 'DeleteActivity', - 'DeleteDataFlowDebugSessionRequest', - 'DelimitedTextDataset', - 'DelimitedTextReadSettings', - 'DelimitedTextSink', - 'DelimitedTextSource', - 'DelimitedTextWriteSettings', - 'DependencyReference', - 'DistcpSettings', - 'DocumentDbCollectionDataset', - 'DocumentDbCollectionSink', - 'DocumentDbCollectionSource', - 'DrillLinkedService', - 'DrillSource', - 'DrillTableDataset', - 'DynamicExecutorAllocation', - 'DynamicsAXLinkedService', - 'DynamicsAXResourceDataset', - 'DynamicsAXSource', - 'DynamicsCrmEntityDataset', - 'DynamicsCrmLinkedService', - 'DynamicsCrmSink', - 'DynamicsCrmSource', - 'DynamicsEntityDataset', - 'DynamicsLinkedService', - 'DynamicsSink', - 'DynamicsSource', - 'EloquaLinkedService', - 'EloquaObjectDataset', - 'EloquaSource', - 'EncryptionDetails', - 'EntityReference', - 'ErrorAdditionalInfo', - 'ErrorContract', - 'ErrorResponse', - 'EvaluateDataFlowExpressionRequest', - 'ExcelDataset', - 'ExcelSource', - 'ExecuteDataFlowActivity', - 'ExecuteDataFlowActivityTypePropertiesCompute', - 'ExecutePipelineActivity', - 'ExecuteSSISPackageActivity', - 'ExecutionActivity', - 'ExportSettings', - 'ExposureControlRequest', - 'ExposureControlResponse', - 'Expression', - 'FileServerLinkedService', - 'FileServerLocation', - 'FileServerReadSettings', - 'FileServerWriteSettings', - 'FileShareDataset', - 'FileSystemSink', - 'FileSystemSource', - 'FilterActivity', - 'ForEachActivity', - 'FormatReadSettings', - 'FormatWriteSettings', - 'FtpReadSettings', - 'FtpServerLinkedService', - 'FtpServerLocation', - 'GetMetadataActivity', - 'GetSsisObjectMetadataRequest', - 'GitHubAccessTokenRequest', - 'GitHubAccessTokenResponse', - 'GitHubClientSecret', - 'GoogleAdWordsLinkedService', - 'GoogleAdWordsObjectDataset', - 'GoogleAdWordsSource', - 'GoogleBigQueryLinkedService', - 'GoogleBigQueryObjectDataset', - 'GoogleBigQuerySource', - 'GoogleCloudStorageLinkedService', - 'GoogleCloudStorageLocation', - 'GoogleCloudStorageReadSettings', - 'GreenplumLinkedService', - 'GreenplumSource', - 'GreenplumTableDataset', - 'HBaseLinkedService', - 'HBaseObjectDataset', - 'HBaseSource', - 'HDInsightHiveActivity', - 'HDInsightLinkedService', - 'HDInsightMapReduceActivity', - 'HDInsightOnDemandLinkedService', - 'HDInsightPigActivity', - 'HDInsightSparkActivity', - 'HDInsightStreamingActivity', - 'HdfsLinkedService', - 'HdfsLocation', - 'HdfsReadSettings', - 'HdfsSource', - 'HiveLinkedService', - 'HiveObjectDataset', - 'HiveSource', - 'HttpDataset', - 'HttpLinkedService', - 'HttpReadSettings', - 'HttpServerLocation', - 'HttpSource', - 'HubspotLinkedService', - 'HubspotObjectDataset', - 'HubspotSource', - 'IfConditionActivity', - 'ImpalaLinkedService', - 'ImpalaObjectDataset', - 'ImpalaSource', - 'ImportSettings', - 'InformixLinkedService', - 'InformixSink', - 'InformixSource', - 'InformixTableDataset', - 'IntegrationRuntime', - 'IntegrationRuntimeComputeProperties', - 'IntegrationRuntimeCustomSetupScriptProperties', - 'IntegrationRuntimeDataFlowProperties', - 'IntegrationRuntimeDataProxyProperties', - 'IntegrationRuntimeListResponse', - 'IntegrationRuntimeReference', - 'IntegrationRuntimeResource', - 'IntegrationRuntimeSsisCatalogInfo', - 'IntegrationRuntimeSsisProperties', - 'IntegrationRuntimeVNetProperties', - 'JiraLinkedService', - 'JiraObjectDataset', - 'JiraSource', - 'JsonDataset', - 'JsonFormat', - 'JsonReadSettings', - 'JsonSink', - 'JsonSource', - 'JsonWriteSettings', - 'LibraryInfo', - 'LibraryListResponse', - 'LibraryRequirements', - 'LibraryResource', - 'LibraryResourceInfo', - 'LibraryResourceProperties', - 'LinkedIntegrationRuntimeKeyAuthorization', - 'LinkedIntegrationRuntimeRbacAuthorization', - 'LinkedIntegrationRuntimeType', - 'LinkedService', - 'LinkedServiceDebugResource', - 'LinkedServiceListResponse', - 'LinkedServiceReference', - 'LinkedServiceResource', - 'LogLocationSettings', - 'LogSettings', - 'LogStorageSettings', - 'LookupActivity', - 'MagentoLinkedService', - 'MagentoObjectDataset', - 'MagentoSource', - 'ManagedIdentity', - 'ManagedIntegrationRuntime', - 'ManagedVirtualNetworkReference', - 'ManagedVirtualNetworkSettings', - 'MappingDataFlow', - 'MariaDBLinkedService', - 'MariaDBSource', - 'MariaDBTableDataset', - 'MarketoLinkedService', - 'MarketoObjectDataset', - 'MarketoSource', - 'MicrosoftAccessLinkedService', - 'MicrosoftAccessSink', - 'MicrosoftAccessSource', - 'MicrosoftAccessTableDataset', - 'MongoDbAtlasCollectionDataset', - 'MongoDbAtlasLinkedService', - 'MongoDbAtlasSource', - 'MongoDbCollectionDataset', - 'MongoDbCursorMethodsProperties', - 'MongoDbLinkedService', - 'MongoDbSource', - 'MongoDbV2CollectionDataset', - 'MongoDbV2LinkedService', - 'MongoDbV2Source', - 'MultiplePipelineTrigger', - 'MySqlLinkedService', - 'MySqlSource', - 'MySqlTableDataset', - 'NetezzaLinkedService', - 'NetezzaPartitionSettings', - 'NetezzaSource', - 'NetezzaTableDataset', - 'Notebook', - 'NotebookCell', - 'NotebookCellOutputItem', - 'NotebookKernelSpec', - 'NotebookLanguageInfo', - 'NotebookListResponse', - 'NotebookMetadata', - 'NotebookResource', - 'NotebookSessionProperties', - 'ODataLinkedService', - 'ODataResourceDataset', - 'ODataSource', - 'OdbcLinkedService', - 'OdbcSink', - 'OdbcSource', - 'OdbcTableDataset', - 'Office365Dataset', - 'Office365LinkedService', - 'Office365Source', - 'OperationResult', - 'OracleLinkedService', - 'OraclePartitionSettings', - 'OracleServiceCloudLinkedService', - 'OracleServiceCloudObjectDataset', - 'OracleServiceCloudSource', - 'OracleSink', - 'OracleSource', - 'OracleTableDataset', - 'OrcDataset', - 'OrcFormat', - 'OrcSink', - 'OrcSource', - 'OrcWriteSettings', - 'ParameterSpecification', - 'ParquetDataset', - 'ParquetFormat', - 'ParquetSink', - 'ParquetSource', - 'ParquetWriteSettings', - 'PaypalLinkedService', - 'PaypalObjectDataset', - 'PaypalSource', - 'PhoenixLinkedService', - 'PhoenixObjectDataset', - 'PhoenixSource', - 'PipelineFolder', - 'PipelineListResponse', - 'PipelineReference', - 'PipelineResource', - 'PipelineRun', - 'PipelineRunInvokedBy', - 'PipelineRunsQueryResponse', - 'PolybaseSettings', - 'PostgreSqlLinkedService', - 'PostgreSqlSource', - 'PostgreSqlTableDataset', - 'PrestoLinkedService', - 'PrestoObjectDataset', - 'PrestoSource', - 'PrivateEndpoint', - 'PrivateEndpointConnection', - 'PrivateLinkServiceConnectionState', - 'ProxyResource', - 'PurviewConfiguration', - 'QueryDataFlowDebugSessionsResponse', - 'QuickBooksLinkedService', - 'QuickBooksObjectDataset', - 'QuickBooksSource', - 'RecurrenceSchedule', - 'RecurrenceScheduleOccurrence', - 'RedirectIncompatibleRowSettings', - 'RedshiftUnloadSettings', - 'RelationalSource', - 'RelationalTableDataset', - 'RerunTriggerListResponse', - 'RerunTriggerResource', - 'RerunTumblingWindowTrigger', - 'RerunTumblingWindowTriggerActionParameters', - 'Resource', - 'ResponsysLinkedService', - 'ResponsysObjectDataset', - 'ResponsysSource', - 'RestResourceDataset', - 'RestServiceLinkedService', - 'RestSink', - 'RestSource', - 'RetryPolicy', - 'RunFilterParameters', - 'RunQueryFilter', - 'RunQueryOrderBy', - 'SSISAccessCredential', - 'SSISChildPackage', - 'SSISExecutionCredential', - 'SSISExecutionParameter', - 'SSISLogLocation', - 'SSISPackageLocation', - 'SSISPropertyOverride', - 'SalesforceLinkedService', - 'SalesforceMarketingCloudLinkedService', - 'SalesforceMarketingCloudObjectDataset', - 'SalesforceMarketingCloudSource', - 'SalesforceObjectDataset', - 'SalesforceServiceCloudLinkedService', - 'SalesforceServiceCloudObjectDataset', - 'SalesforceServiceCloudSink', - 'SalesforceServiceCloudSource', - 'SalesforceSink', - 'SalesforceSource', - 'SapBWLinkedService', - 'SapBwCubeDataset', - 'SapBwSource', - 'SapCloudForCustomerLinkedService', - 'SapCloudForCustomerResourceDataset', - 'SapCloudForCustomerSink', - 'SapCloudForCustomerSource', - 'SapEccLinkedService', - 'SapEccResourceDataset', - 'SapEccSource', - 'SapHanaLinkedService', - 'SapHanaPartitionSettings', - 'SapHanaSource', - 'SapHanaTableDataset', - 'SapOpenHubLinkedService', - 'SapOpenHubSource', - 'SapOpenHubTableDataset', - 'SapTableLinkedService', - 'SapTablePartitionSettings', - 'SapTableResourceDataset', - 'SapTableSource', - 'ScheduleTrigger', - 'ScheduleTriggerRecurrence', - 'ScriptAction', - 'SecretBase', - 'SecureString', - 'SelfDependencyTumblingWindowTriggerReference', - 'SelfHostedIntegrationRuntime', - 'ServiceNowLinkedService', - 'ServiceNowObjectDataset', - 'ServiceNowSource', - 'SetVariableActivity', - 'SftpLocation', - 'SftpReadSettings', - 'SftpServerLinkedService', - 'SftpWriteSettings', - 'SharePointOnlineListLinkedService', - 'SharePointOnlineListResourceDataset', - 'SharePointOnlineListSource', - 'ShopifyLinkedService', - 'ShopifyObjectDataset', - 'ShopifySource', - 'SkipErrorFile', - 'Sku', - 'SnowflakeDataset', - 'SnowflakeExportCopyCommand', - 'SnowflakeImportCopyCommand', - 'SnowflakeLinkedService', - 'SnowflakeSink', - 'SnowflakeSource', - 'SparkBatchJob', - 'SparkBatchJobState', - 'SparkJobDefinition', - 'SparkJobDefinitionResource', - 'SparkJobDefinitionsListResponse', - 'SparkJobProperties', - 'SparkLinkedService', - 'SparkObjectDataset', - 'SparkRequest', - 'SparkScheduler', - 'SparkServiceError', - 'SparkServicePlugin', - 'SparkSource', - 'SqlConnection', - 'SqlDWSink', - 'SqlDWSource', - 'SqlMISink', - 'SqlMISource', - 'SqlPartitionSettings', - 'SqlPool', - 'SqlPoolInfoListResult', - 'SqlPoolReference', - 'SqlPoolStoredProcedureActivity', - 'SqlScript', - 'SqlScriptContent', - 'SqlScriptMetadata', - 'SqlScriptResource', - 'SqlScriptsListResponse', - 'SqlServerLinkedService', - 'SqlServerSink', - 'SqlServerSource', - 'SqlServerStoredProcedureActivity', - 'SqlServerTableDataset', - 'SqlSink', - 'SqlSource', - 'SquareLinkedService', - 'SquareObjectDataset', - 'SquareSource', - 'SsisObjectMetadataStatusResponse', - 'StagingSettings', - 'StartDataFlowDebugSessionRequest', - 'StartDataFlowDebugSessionResponse', - 'StoreReadSettings', - 'StoreWriteSettings', - 'StoredProcedureParameter', - 'SubResource', - 'SubResourceDebugResource', - 'SwitchActivity', - 'SwitchCase', - 'SybaseLinkedService', - 'SybaseSource', - 'SybaseTableDataset', - 'SynapseNotebookActivity', - 'SynapseNotebookReference', - 'SynapseSparkJobDefinitionActivity', - 'SynapseSparkJobReference', - 'TabularSource', - 'TabularTranslator', - 'TarGZipReadSettings', - 'TarReadSettings', - 'TeradataLinkedService', - 'TeradataPartitionSettings', - 'TeradataSource', - 'TeradataTableDataset', - 'TextFormat', - 'TrackedResource', - 'Transformation', - 'Trigger', - 'TriggerDependencyProvisioningStatus', - 'TriggerDependencyReference', - 'TriggerListResponse', - 'TriggerPipelineReference', - 'TriggerReference', - 'TriggerResource', - 'TriggerRun', - 'TriggerRunsQueryResponse', - 'TriggerSubscriptionOperationStatus', - 'TumblingWindowTrigger', - 'TumblingWindowTriggerDependencyReference', - 'TypeConversionSettings', - 'UntilActivity', - 'UserProperty', - 'ValidationActivity', - 'VariableSpecification', - 'VerticaLinkedService', - 'VerticaSource', - 'VerticaTableDataset', - 'VirtualNetworkProfile', - 'WaitActivity', - 'WebActivity', - 'WebActivityAuthentication', - 'WebAnonymousAuthentication', - 'WebBasicAuthentication', - 'WebClientCertificateAuthentication', - 'WebHookActivity', - 'WebLinkedService', - 'WebLinkedServiceTypeProperties', - 'WebSource', - 'WebTableDataset', - 'Workspace', - 'WorkspaceIdentity', - 'WorkspaceKeyDetails', - 'WorkspaceRepositoryConfiguration', - 'WorkspaceUpdateParameters', - 'XeroLinkedService', - 'XeroObjectDataset', - 'XeroSource', - 'XmlDataset', - 'XmlReadSettings', - 'XmlSource', - 'ZipDeflateReadSettings', - 'ZohoLinkedService', - 'ZohoObjectDataset', - 'ZohoSource', - 'AvroCompressionCodec', - 'AzureFunctionActivityMethod', - 'AzureSearchIndexWriteBehaviorType', - 'BigDataPoolReferenceType', - 'BlobEventType', - 'CassandraSourceReadConsistencyLevels', - 'CellOutputType', - 'CompressionCodec', - 'CopyBehaviorType', - 'DataFlowComputeType', - 'DataFlowReferenceType', - 'DatasetCompressionLevel', - 'DatasetReferenceType', - 'DayOfWeek', - 'Db2AuthenticationType', - 'DependencyCondition', - 'DynamicsAuthenticationType', - 'DynamicsDeploymentType', - 'DynamicsServicePrincipalCredentialType', - 'DynamicsSinkWriteBehavior', - 'EventSubscriptionStatus', - 'ExpressionType', - 'FtpAuthenticationType', - 'GoogleAdWordsAuthenticationType', - 'GoogleBigQueryAuthenticationType', - 'HBaseAuthenticationType', - 'HDInsightActivityDebugInfoOption', - 'HdiNodeTypes', - 'HiveAuthenticationType', - 'HiveServerType', - 'HiveThriftTransportProtocol', - 'HttpAuthenticationType', - 'ImpalaAuthenticationType', - 'IntegrationRuntimeEdition', - 'IntegrationRuntimeEntityReferenceType', - 'IntegrationRuntimeLicenseType', - 'IntegrationRuntimeReferenceType', - 'IntegrationRuntimeSsisCatalogPricingTier', - 'IntegrationRuntimeState', - 'IntegrationRuntimeType', - 'JsonFormatFilePattern', - 'JsonWriteFilePattern', - 'MongoDbAuthenticationType', - 'NetezzaPartitionOption', - 'NodeSize', - 'NodeSizeFamily', - 'NotebookReferenceType', - 'ODataAadServicePrincipalCredentialType', - 'ODataAuthenticationType', - 'OraclePartitionOption', - 'OrcCompressionCodec', - 'ParameterType', - 'ParquetCompressionCodecEnum', - 'PhoenixAuthenticationType', - 'PipelineReferenceType', - 'PluginCurrentState', - 'PolybaseSettingsRejectType', - 'PrestoAuthenticationType', - 'RecurrenceFrequency', - 'ResourceIdentityType', - 'RestServiceAuthenticationType', - 'RunQueryFilterOperand', - 'RunQueryFilterOperator', - 'RunQueryOrder', - 'RunQueryOrderByField', - 'SalesforceSinkWriteBehavior', - 'SalesforceSourceReadBehavior', - 'SapCloudForCustomerSinkWriteBehavior', - 'SapHanaAuthenticationType', - 'SapHanaPartitionOption', - 'SapTablePartitionOption', - 'SchedulerCurrentState', - 'ServiceNowAuthenticationType', - 'SftpAuthenticationType', - 'SparkAuthenticationType', - 'SparkBatchJobResultType', - 'SparkErrorSource', - 'SparkJobReferenceType', - 'SparkJobType', - 'SparkServerType', - 'SparkThriftTransportProtocol', - 'SqlConnectionType', - 'SqlPartitionOption', - 'SqlPoolReferenceType', - 'SqlScriptType', - 'SsisLogLocationType', - 'SsisPackageLocationType', - 'StoredProcedureParameterType', - 'SybaseAuthenticationType', - 'TeradataAuthenticationType', - 'TeradataPartitionOption', - 'TriggerReferenceType', - 'TriggerRunStatus', - 'TriggerRuntimeState', - 'TumblingWindowFrequency', - 'Type', - 'VariableType', - 'WebActivityMethod', - 'WebAuthenticationType', - 'WebHookActivityMethod', -] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/models/_models.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/models/_models.py deleted file mode 100644 index 423327d46c23..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/models/_models.py +++ /dev/null @@ -1,36083 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from azure.core.exceptions import HttpResponseError -import msrest.serialization - - -class Activity(msrest.serialization.Model): - """A pipeline activity. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: ControlActivity, ExecutionActivity, SqlPoolStoredProcedureActivity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - } - - _subtype_map = { - 'type': {'Container': 'ControlActivity', 'Execution': 'ExecutionActivity', 'SqlPoolStoredProcedure': 'SqlPoolStoredProcedureActivity'} - } - - def __init__( - self, - **kwargs - ): - super(Activity, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.name = kwargs['name'] - self.type = 'Activity' # type: str - self.description = kwargs.get('description', None) - self.depends_on = kwargs.get('depends_on', None) - self.user_properties = kwargs.get('user_properties', None) - - -class ActivityDependency(msrest.serialization.Model): - """Activity dependency information. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param activity: Required. Activity name. - :type activity: str - :param dependency_conditions: Required. Match-Condition for the dependency. - :type dependency_conditions: list[str or - ~azure.synapse.artifacts.v2020_12_01.models.DependencyCondition] - """ - - _validation = { - 'activity': {'required': True}, - 'dependency_conditions': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'activity': {'key': 'activity', 'type': 'str'}, - 'dependency_conditions': {'key': 'dependencyConditions', 'type': '[str]'}, - } - - def __init__( - self, - **kwargs - ): - super(ActivityDependency, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.activity = kwargs['activity'] - self.dependency_conditions = kwargs['dependency_conditions'] - - -class ActivityPolicy(msrest.serialization.Model): - """Execution policy for an activity. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param timeout: Specifies the timeout for the activity to run. The default timeout is 7 days. - Type: string (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type timeout: any - :param retry: Maximum ordinary retry attempts. Default is 0. Type: integer (or Expression with - resultType integer), minimum: 0. - :type retry: any - :param retry_interval_in_seconds: Interval between each retry attempt (in seconds). The default - is 30 sec. - :type retry_interval_in_seconds: int - :param secure_input: When set to true, Input from activity is considered as secure and will not - be logged to monitoring. - :type secure_input: bool - :param secure_output: When set to true, Output from activity is considered as secure and will - not be logged to monitoring. - :type secure_output: bool - """ - - _validation = { - 'retry_interval_in_seconds': {'maximum': 86400, 'minimum': 30}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'timeout': {'key': 'timeout', 'type': 'object'}, - 'retry': {'key': 'retry', 'type': 'object'}, - 'retry_interval_in_seconds': {'key': 'retryIntervalInSeconds', 'type': 'int'}, - 'secure_input': {'key': 'secureInput', 'type': 'bool'}, - 'secure_output': {'key': 'secureOutput', 'type': 'bool'}, - } - - def __init__( - self, - **kwargs - ): - super(ActivityPolicy, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.timeout = kwargs.get('timeout', None) - self.retry = kwargs.get('retry', None) - self.retry_interval_in_seconds = kwargs.get('retry_interval_in_seconds', None) - self.secure_input = kwargs.get('secure_input', None) - self.secure_output = kwargs.get('secure_output', None) - - -class ActivityRun(msrest.serialization.Model): - """Information about an activity run in a pipeline. - - Variables are only populated by the server, and will be ignored when sending a request. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :ivar pipeline_name: The name of the pipeline. - :vartype pipeline_name: str - :ivar pipeline_run_id: The id of the pipeline run. - :vartype pipeline_run_id: str - :ivar activity_name: The name of the activity. - :vartype activity_name: str - :ivar activity_type: The type of the activity. - :vartype activity_type: str - :ivar activity_run_id: The id of the activity run. - :vartype activity_run_id: str - :ivar linked_service_name: The name of the compute linked service. - :vartype linked_service_name: str - :ivar status: The status of the activity run. - :vartype status: str - :ivar activity_run_start: The start time of the activity run in 'ISO 8601' format. - :vartype activity_run_start: ~datetime.datetime - :ivar activity_run_end: The end time of the activity run in 'ISO 8601' format. - :vartype activity_run_end: ~datetime.datetime - :ivar duration_in_ms: The duration of the activity run. - :vartype duration_in_ms: int - :ivar input: The input for the activity. - :vartype input: any - :ivar output: The output for the activity. - :vartype output: any - :ivar error: The error if any from the activity run. - :vartype error: any - """ - - _validation = { - 'pipeline_name': {'readonly': True}, - 'pipeline_run_id': {'readonly': True}, - 'activity_name': {'readonly': True}, - 'activity_type': {'readonly': True}, - 'activity_run_id': {'readonly': True}, - 'linked_service_name': {'readonly': True}, - 'status': {'readonly': True}, - 'activity_run_start': {'readonly': True}, - 'activity_run_end': {'readonly': True}, - 'duration_in_ms': {'readonly': True}, - 'input': {'readonly': True}, - 'output': {'readonly': True}, - 'error': {'readonly': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'pipeline_name': {'key': 'pipelineName', 'type': 'str'}, - 'pipeline_run_id': {'key': 'pipelineRunId', 'type': 'str'}, - 'activity_name': {'key': 'activityName', 'type': 'str'}, - 'activity_type': {'key': 'activityType', 'type': 'str'}, - 'activity_run_id': {'key': 'activityRunId', 'type': 'str'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'str'}, - 'status': {'key': 'status', 'type': 'str'}, - 'activity_run_start': {'key': 'activityRunStart', 'type': 'iso-8601'}, - 'activity_run_end': {'key': 'activityRunEnd', 'type': 'iso-8601'}, - 'duration_in_ms': {'key': 'durationInMs', 'type': 'int'}, - 'input': {'key': 'input', 'type': 'object'}, - 'output': {'key': 'output', 'type': 'object'}, - 'error': {'key': 'error', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ActivityRun, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.pipeline_name = None - self.pipeline_run_id = None - self.activity_name = None - self.activity_type = None - self.activity_run_id = None - self.linked_service_name = None - self.status = None - self.activity_run_start = None - self.activity_run_end = None - self.duration_in_ms = None - self.input = None - self.output = None - self.error = None - - -class ActivityRunsQueryResponse(msrest.serialization.Model): - """A list activity runs. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of activity runs. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityRun] - :param continuation_token: The continuation token for getting the next page of results, if any - remaining results exist, null otherwise. - :type continuation_token: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[ActivityRun]'}, - 'continuation_token': {'key': 'continuationToken', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(ActivityRunsQueryResponse, self).__init__(**kwargs) - self.value = kwargs['value'] - self.continuation_token = kwargs.get('continuation_token', None) - - -class AddDataFlowToDebugSessionResponse(msrest.serialization.Model): - """Response body structure for starting data flow debug session. - - :param job_version: The ID of data flow debug job version. - :type job_version: str - """ - - _attribute_map = { - 'job_version': {'key': 'jobVersion', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(AddDataFlowToDebugSessionResponse, self).__init__(**kwargs) - self.job_version = kwargs.get('job_version', None) - - -class AdditionalColumns(msrest.serialization.Model): - """Specify the column name and value of additional columns. - - :param name: Additional column name. Type: string (or Expression with resultType string). - :type name: any - :param value: Additional column value. Type: string (or Expression with resultType string). - :type value: any - """ - - _attribute_map = { - 'name': {'key': 'name', 'type': 'object'}, - 'value': {'key': 'value', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AdditionalColumns, self).__init__(**kwargs) - self.name = kwargs.get('name', None) - self.value = kwargs.get('value', None) - - -class LinkedService(msrest.serialization.Model): - """The Azure Synapse nested object which contains the information and credential which can be used to connect with related store or compute resource. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonMWSLinkedService, AmazonRedshiftLinkedService, AmazonS3LinkedService, AzureBatchLinkedService, AzureBlobFSLinkedService, AzureBlobStorageLinkedService, AzureDataExplorerLinkedService, AzureDataLakeAnalyticsLinkedService, AzureDataLakeStoreLinkedService, AzureDatabricksLinkedService, AzureDatabricksDeltaLakeLinkedService, AzureFileStorageLinkedService, AzureFunctionLinkedService, AzureKeyVaultLinkedService, AzureMLLinkedService, AzureMLServiceLinkedService, AzureMariaDBLinkedService, AzureMySqlLinkedService, AzurePostgreSqlLinkedService, AzureSearchLinkedService, AzureSqlDWLinkedService, AzureSqlDatabaseLinkedService, AzureSqlMILinkedService, AzureStorageLinkedService, AzureTableStorageLinkedService, CassandraLinkedService, CommonDataServiceForAppsLinkedService, ConcurLinkedService, CosmosDbLinkedService, CosmosDbMongoDbApiLinkedService, CouchbaseLinkedService, CustomDataSourceLinkedService, Db2LinkedService, DrillLinkedService, DynamicsLinkedService, DynamicsAXLinkedService, DynamicsCrmLinkedService, EloquaLinkedService, FileServerLinkedService, FtpServerLinkedService, GoogleAdWordsLinkedService, GoogleBigQueryLinkedService, GoogleCloudStorageLinkedService, GreenplumLinkedService, HBaseLinkedService, HDInsightLinkedService, HDInsightOnDemandLinkedService, HdfsLinkedService, HiveLinkedService, HttpLinkedService, HubspotLinkedService, ImpalaLinkedService, InformixLinkedService, JiraLinkedService, MagentoLinkedService, MariaDBLinkedService, MarketoLinkedService, MicrosoftAccessLinkedService, MongoDbLinkedService, MongoDbAtlasLinkedService, MongoDbV2LinkedService, MySqlLinkedService, NetezzaLinkedService, ODataLinkedService, OdbcLinkedService, Office365LinkedService, OracleLinkedService, OracleServiceCloudLinkedService, PaypalLinkedService, PhoenixLinkedService, PostgreSqlLinkedService, PrestoLinkedService, QuickBooksLinkedService, ResponsysLinkedService, RestServiceLinkedService, SalesforceLinkedService, SalesforceMarketingCloudLinkedService, SalesforceServiceCloudLinkedService, SapBWLinkedService, SapCloudForCustomerLinkedService, SapEccLinkedService, SapHanaLinkedService, SapOpenHubLinkedService, SapTableLinkedService, ServiceNowLinkedService, SftpServerLinkedService, SharePointOnlineListLinkedService, ShopifyLinkedService, SnowflakeLinkedService, SparkLinkedService, SqlServerLinkedService, SquareLinkedService, SybaseLinkedService, TeradataLinkedService, VerticaLinkedService, WebLinkedService, XeroLinkedService, ZohoLinkedService. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - } - - _subtype_map = { - 'type': {'AmazonMWS': 'AmazonMWSLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureBlobFS': 'AzureBlobFSLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureDataExplorer': 'AzureDataExplorerLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDatabricksDeltaLake': 'AzureDatabricksDeltaLakeLinkedService', 'AzureFileStorage': 'AzureFileStorageLinkedService', 'AzureFunction': 'AzureFunctionLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureML': 'AzureMLLinkedService', 'AzureMLService': 'AzureMLServiceLinkedService', 'AzureMariaDB': 'AzureMariaDBLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'AzureSqlMI': 'AzureSqlMILinkedService', 'AzureStorage': 'AzureStorageLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'Cassandra': 'CassandraLinkedService', 'CommonDataServiceForApps': 'CommonDataServiceForAppsLinkedService', 'Concur': 'ConcurLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'CosmosDbMongoDbApi': 'CosmosDbMongoDbApiLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'Db2': 'Db2LinkedService', 'Drill': 'DrillLinkedService', 'Dynamics': 'DynamicsLinkedService', 'DynamicsAX': 'DynamicsAXLinkedService', 'DynamicsCrm': 'DynamicsCrmLinkedService', 'Eloqua': 'EloquaLinkedService', 'FileServer': 'FileServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'GoogleAdWords': 'GoogleAdWordsLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'GoogleCloudStorage': 'GoogleCloudStorageLinkedService', 'Greenplum': 'GreenplumLinkedService', 'HBase': 'HBaseLinkedService', 'HDInsight': 'HDInsightLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'Hdfs': 'HdfsLinkedService', 'Hive': 'HiveLinkedService', 'HttpServer': 'HttpLinkedService', 'Hubspot': 'HubspotLinkedService', 'Impala': 'ImpalaLinkedService', 'Informix': 'InformixLinkedService', 'Jira': 'JiraLinkedService', 'Magento': 'MagentoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Marketo': 'MarketoLinkedService', 'MicrosoftAccess': 'MicrosoftAccessLinkedService', 'MongoDb': 'MongoDbLinkedService', 'MongoDbAtlas': 'MongoDbAtlasLinkedService', 'MongoDbV2': 'MongoDbV2LinkedService', 'MySql': 'MySqlLinkedService', 'Netezza': 'NetezzaLinkedService', 'OData': 'ODataLinkedService', 'Odbc': 'OdbcLinkedService', 'Office365': 'Office365LinkedService', 'Oracle': 'OracleLinkedService', 'OracleServiceCloud': 'OracleServiceCloudLinkedService', 'Paypal': 'PaypalLinkedService', 'Phoenix': 'PhoenixLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'Presto': 'PrestoLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Responsys': 'ResponsysLinkedService', 'RestService': 'RestServiceLinkedService', 'Salesforce': 'SalesforceLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'SalesforceServiceCloud': 'SalesforceServiceCloudLinkedService', 'SapBW': 'SapBWLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'SapEcc': 'SapEccLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapOpenHub': 'SapOpenHubLinkedService', 'SapTable': 'SapTableLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'Sftp': 'SftpServerLinkedService', 'SharePointOnlineList': 'SharePointOnlineListLinkedService', 'Shopify': 'ShopifyLinkedService', 'Snowflake': 'SnowflakeLinkedService', 'Spark': 'SparkLinkedService', 'SqlServer': 'SqlServerLinkedService', 'Square': 'SquareLinkedService', 'Sybase': 'SybaseLinkedService', 'Teradata': 'TeradataLinkedService', 'Vertica': 'VerticaLinkedService', 'Web': 'WebLinkedService', 'Xero': 'XeroLinkedService', 'Zoho': 'ZohoLinkedService'} - } - - def __init__( - self, - **kwargs - ): - super(LinkedService, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = 'LinkedService' # type: str - self.connect_via = kwargs.get('connect_via', None) - self.description = kwargs.get('description', None) - self.parameters = kwargs.get('parameters', None) - self.annotations = kwargs.get('annotations', None) - - -class AmazonMWSLinkedService(LinkedService): - """Amazon Marketplace Web Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param endpoint: Required. The endpoint of the Amazon MWS server, (i.e. - mws.amazonservices.com). - :type endpoint: any - :param marketplace_id: Required. The Amazon Marketplace ID you want to retrieve data from. To - retrieve data from multiple Marketplace IDs, separate them with a comma (,). (i.e. - A2EUQ1WTGCTBG2). - :type marketplace_id: any - :param seller_id: Required. The Amazon seller ID. - :type seller_id: any - :param mws_auth_token: The Amazon MWS authentication token. - :type mws_auth_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param access_key_id: Required. The access key id used to access data. - :type access_key_id: any - :param secret_key: The secret key used to access data. - :type secret_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'endpoint': {'required': True}, - 'marketplace_id': {'required': True}, - 'seller_id': {'required': True}, - 'access_key_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, - 'marketplace_id': {'key': 'typeProperties.marketplaceID', 'type': 'object'}, - 'seller_id': {'key': 'typeProperties.sellerID', 'type': 'object'}, - 'mws_auth_token': {'key': 'typeProperties.mwsAuthToken', 'type': 'SecretBase'}, - 'access_key_id': {'key': 'typeProperties.accessKeyId', 'type': 'object'}, - 'secret_key': {'key': 'typeProperties.secretKey', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AmazonMWSLinkedService, self).__init__(**kwargs) - self.type = 'AmazonMWS' # type: str - self.endpoint = kwargs['endpoint'] - self.marketplace_id = kwargs['marketplace_id'] - self.seller_id = kwargs['seller_id'] - self.mws_auth_token = kwargs.get('mws_auth_token', None) - self.access_key_id = kwargs['access_key_id'] - self.secret_key = kwargs.get('secret_key', None) - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.use_host_verification = kwargs.get('use_host_verification', None) - self.use_peer_verification = kwargs.get('use_peer_verification', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class Dataset(msrest.serialization.Model): - """The Azure Data Factory nested object which identifies data within different data stores, such as tables, files, folders, and documents. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonMWSObjectDataset, AmazonRedshiftTableDataset, AmazonS3Dataset, AvroDataset, AzureBlobDataset, AzureBlobFSDataset, AzureDataExplorerTableDataset, AzureDataLakeStoreDataset, AzureDatabricksDeltaLakeDataset, AzureMariaDBTableDataset, AzureMySqlTableDataset, AzurePostgreSqlTableDataset, AzureSearchIndexDataset, AzureSqlDWTableDataset, AzureSqlMITableDataset, AzureSqlTableDataset, AzureTableDataset, BinaryDataset, CassandraTableDataset, CommonDataServiceForAppsEntityDataset, ConcurObjectDataset, CosmosDbMongoDbApiCollectionDataset, CosmosDbSqlApiCollectionDataset, CouchbaseTableDataset, CustomDataset, Db2TableDataset, DelimitedTextDataset, DocumentDbCollectionDataset, DrillTableDataset, DynamicsAXResourceDataset, DynamicsCrmEntityDataset, DynamicsEntityDataset, EloquaObjectDataset, ExcelDataset, FileShareDataset, GoogleAdWordsObjectDataset, GoogleBigQueryObjectDataset, GreenplumTableDataset, HBaseObjectDataset, HiveObjectDataset, HttpDataset, HubspotObjectDataset, ImpalaObjectDataset, InformixTableDataset, JiraObjectDataset, JsonDataset, MagentoObjectDataset, MariaDBTableDataset, MarketoObjectDataset, MicrosoftAccessTableDataset, MongoDbAtlasCollectionDataset, MongoDbCollectionDataset, MongoDbV2CollectionDataset, MySqlTableDataset, NetezzaTableDataset, ODataResourceDataset, OdbcTableDataset, Office365Dataset, OracleServiceCloudObjectDataset, OracleTableDataset, OrcDataset, ParquetDataset, PaypalObjectDataset, PhoenixObjectDataset, PostgreSqlTableDataset, PrestoObjectDataset, QuickBooksObjectDataset, RelationalTableDataset, ResponsysObjectDataset, RestResourceDataset, SalesforceMarketingCloudObjectDataset, SalesforceObjectDataset, SalesforceServiceCloudObjectDataset, SapBwCubeDataset, SapCloudForCustomerResourceDataset, SapEccResourceDataset, SapHanaTableDataset, SapOpenHubTableDataset, SapTableResourceDataset, ServiceNowObjectDataset, SharePointOnlineListResourceDataset, ShopifyObjectDataset, SnowflakeDataset, SparkObjectDataset, SqlServerTableDataset, SquareObjectDataset, SybaseTableDataset, TeradataTableDataset, VerticaTableDataset, WebTableDataset, XeroObjectDataset, XmlDataset, ZohoObjectDataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - } - - _subtype_map = { - 'type': {'AmazonMWSObject': 'AmazonMWSObjectDataset', 'AmazonRedshiftTable': 'AmazonRedshiftTableDataset', 'AmazonS3Object': 'AmazonS3Dataset', 'Avro': 'AvroDataset', 'AzureBlob': 'AzureBlobDataset', 'AzureBlobFSFile': 'AzureBlobFSDataset', 'AzureDataExplorerTable': 'AzureDataExplorerTableDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'AzureDatabricksDeltaLakeDataset': 'AzureDatabricksDeltaLakeDataset', 'AzureMariaDBTable': 'AzureMariaDBTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlMITable': 'AzureSqlMITableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'Binary': 'BinaryDataset', 'CassandraTable': 'CassandraTableDataset', 'CommonDataServiceForAppsEntity': 'CommonDataServiceForAppsEntityDataset', 'ConcurObject': 'ConcurObjectDataset', 'CosmosDbMongoDbApiCollection': 'CosmosDbMongoDbApiCollectionDataset', 'CosmosDbSqlApiCollection': 'CosmosDbSqlApiCollectionDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'CustomDataset': 'CustomDataset', 'Db2Table': 'Db2TableDataset', 'DelimitedText': 'DelimitedTextDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'DrillTable': 'DrillTableDataset', 'DynamicsAXResource': 'DynamicsAXResourceDataset', 'DynamicsCrmEntity': 'DynamicsCrmEntityDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'EloquaObject': 'EloquaObjectDataset', 'Excel': 'ExcelDataset', 'FileShare': 'FileShareDataset', 'GoogleAdWordsObject': 'GoogleAdWordsObjectDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'HBaseObject': 'HBaseObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HttpFile': 'HttpDataset', 'HubspotObject': 'HubspotObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'InformixTable': 'InformixTableDataset', 'JiraObject': 'JiraObjectDataset', 'Json': 'JsonDataset', 'MagentoObject': 'MagentoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MarketoObject': 'MarketoObjectDataset', 'MicrosoftAccessTable': 'MicrosoftAccessTableDataset', 'MongoDbAtlasCollection': 'MongoDbAtlasCollectionDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'MongoDbV2Collection': 'MongoDbV2CollectionDataset', 'MySqlTable': 'MySqlTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ODataResource': 'ODataResourceDataset', 'OdbcTable': 'OdbcTableDataset', 'Office365Table': 'Office365Dataset', 'OracleServiceCloudObject': 'OracleServiceCloudObjectDataset', 'OracleTable': 'OracleTableDataset', 'Orc': 'OrcDataset', 'Parquet': 'ParquetDataset', 'PaypalObject': 'PaypalObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PostgreSqlTable': 'PostgreSqlTableDataset', 'PrestoObject': 'PrestoObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'ResponsysObject': 'ResponsysObjectDataset', 'RestResource': 'RestResourceDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'SalesforceServiceCloudObject': 'SalesforceServiceCloudObjectDataset', 'SapBwCube': 'SapBwCubeDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapHanaTable': 'SapHanaTableDataset', 'SapOpenHubTable': 'SapOpenHubTableDataset', 'SapTableResource': 'SapTableResourceDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'SharePointOnlineListResource': 'SharePointOnlineListResourceDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'SnowflakeTable': 'SnowflakeDataset', 'SparkObject': 'SparkObjectDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SquareObject': 'SquareObjectDataset', 'SybaseTable': 'SybaseTableDataset', 'TeradataTable': 'TeradataTableDataset', 'VerticaTable': 'VerticaTableDataset', 'WebTable': 'WebTableDataset', 'XeroObject': 'XeroObjectDataset', 'Xml': 'XmlDataset', 'ZohoObject': 'ZohoObjectDataset'} - } - - def __init__( - self, - **kwargs - ): - super(Dataset, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = 'Dataset' # type: str - self.description = kwargs.get('description', None) - self.structure = kwargs.get('structure', None) - self.schema = kwargs.get('schema', None) - self.linked_service_name = kwargs['linked_service_name'] - self.parameters = kwargs.get('parameters', None) - self.annotations = kwargs.get('annotations', None) - self.folder = kwargs.get('folder', None) - - -class AmazonMWSObjectDataset(Dataset): - """Amazon Marketplace Web Service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AmazonMWSObjectDataset, self).__init__(**kwargs) - self.type = 'AmazonMWSObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class CopySource(msrest.serialization.Model): - """A copy activity source. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AvroSource, AzureBlobFSSource, AzureDataExplorerSource, AzureDataLakeStoreSource, AzureDatabricksDeltaLakeSource, BinarySource, BlobSource, CommonDataServiceForAppsSource, CosmosDbMongoDbApiSource, CosmosDbSqlApiSource, DelimitedTextSource, DocumentDbCollectionSource, DynamicsCrmSource, DynamicsSource, ExcelSource, FileSystemSource, HdfsSource, HttpSource, JsonSource, MicrosoftAccessSource, MongoDbAtlasSource, MongoDbSource, MongoDbV2Source, ODataSource, Office365Source, OracleSource, OrcSource, ParquetSource, RelationalSource, RestSource, SalesforceServiceCloudSource, SharePointOnlineListSource, SnowflakeSource, TabularSource, WebSource, XmlSource. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - } - - _subtype_map = { - 'type': {'AvroSource': 'AvroSource', 'AzureBlobFSSource': 'AzureBlobFSSource', 'AzureDataExplorerSource': 'AzureDataExplorerSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'AzureDatabricksDeltaLakeSource': 'AzureDatabricksDeltaLakeSource', 'BinarySource': 'BinarySource', 'BlobSource': 'BlobSource', 'CommonDataServiceForAppsSource': 'CommonDataServiceForAppsSource', 'CosmosDbMongoDbApiSource': 'CosmosDbMongoDbApiSource', 'CosmosDbSqlApiSource': 'CosmosDbSqlApiSource', 'DelimitedTextSource': 'DelimitedTextSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'DynamicsCrmSource': 'DynamicsCrmSource', 'DynamicsSource': 'DynamicsSource', 'ExcelSource': 'ExcelSource', 'FileSystemSource': 'FileSystemSource', 'HdfsSource': 'HdfsSource', 'HttpSource': 'HttpSource', 'JsonSource': 'JsonSource', 'MicrosoftAccessSource': 'MicrosoftAccessSource', 'MongoDbAtlasSource': 'MongoDbAtlasSource', 'MongoDbSource': 'MongoDbSource', 'MongoDbV2Source': 'MongoDbV2Source', 'ODataSource': 'ODataSource', 'Office365Source': 'Office365Source', 'OracleSource': 'OracleSource', 'OrcSource': 'OrcSource', 'ParquetSource': 'ParquetSource', 'RelationalSource': 'RelationalSource', 'RestSource': 'RestSource', 'SalesforceServiceCloudSource': 'SalesforceServiceCloudSource', 'SharePointOnlineListSource': 'SharePointOnlineListSource', 'SnowflakeSource': 'SnowflakeSource', 'TabularSource': 'TabularSource', 'WebSource': 'WebSource', 'XmlSource': 'XmlSource'} - } - - def __init__( - self, - **kwargs - ): - super(CopySource, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = 'CopySource' # type: str - self.source_retry_count = kwargs.get('source_retry_count', None) - self.source_retry_wait = kwargs.get('source_retry_wait', None) - self.max_concurrent_connections = kwargs.get('max_concurrent_connections', None) - - -class TabularSource(CopySource): - """Copy activity sources of tabular type. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonMWSSource, AmazonRedshiftSource, AzureMariaDBSource, AzureMySqlSource, AzurePostgreSqlSource, AzureSqlSource, AzureTableSource, CassandraSource, ConcurSource, CouchbaseSource, Db2Source, DrillSource, DynamicsAXSource, EloquaSource, GoogleAdWordsSource, GoogleBigQuerySource, GreenplumSource, HBaseSource, HiveSource, HubspotSource, ImpalaSource, InformixSource, JiraSource, MagentoSource, MariaDBSource, MarketoSource, MySqlSource, NetezzaSource, OdbcSource, OracleServiceCloudSource, PaypalSource, PhoenixSource, PostgreSqlSource, PrestoSource, QuickBooksSource, ResponsysSource, SalesforceMarketingCloudSource, SalesforceSource, SapBwSource, SapCloudForCustomerSource, SapEccSource, SapHanaSource, SapOpenHubSource, SapTableSource, ServiceNowSource, ShopifySource, SparkSource, SqlDWSource, SqlMISource, SqlServerSource, SqlSource, SquareSource, SybaseSource, TeradataSource, VerticaSource, XeroSource, ZohoSource. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - _subtype_map = { - 'type': {'AmazonMWSSource': 'AmazonMWSSource', 'AmazonRedshiftSource': 'AmazonRedshiftSource', 'AzureMariaDBSource': 'AzureMariaDBSource', 'AzureMySqlSource': 'AzureMySqlSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AzureSqlSource': 'AzureSqlSource', 'AzureTableSource': 'AzureTableSource', 'CassandraSource': 'CassandraSource', 'ConcurSource': 'ConcurSource', 'CouchbaseSource': 'CouchbaseSource', 'Db2Source': 'Db2Source', 'DrillSource': 'DrillSource', 'DynamicsAXSource': 'DynamicsAXSource', 'EloquaSource': 'EloquaSource', 'GoogleAdWordsSource': 'GoogleAdWordsSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'GreenplumSource': 'GreenplumSource', 'HBaseSource': 'HBaseSource', 'HiveSource': 'HiveSource', 'HubspotSource': 'HubspotSource', 'ImpalaSource': 'ImpalaSource', 'InformixSource': 'InformixSource', 'JiraSource': 'JiraSource', 'MagentoSource': 'MagentoSource', 'MariaDBSource': 'MariaDBSource', 'MarketoSource': 'MarketoSource', 'MySqlSource': 'MySqlSource', 'NetezzaSource': 'NetezzaSource', 'OdbcSource': 'OdbcSource', 'OracleServiceCloudSource': 'OracleServiceCloudSource', 'PaypalSource': 'PaypalSource', 'PhoenixSource': 'PhoenixSource', 'PostgreSqlSource': 'PostgreSqlSource', 'PrestoSource': 'PrestoSource', 'QuickBooksSource': 'QuickBooksSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'SalesforceSource': 'SalesforceSource', 'SapBwSource': 'SapBwSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SapEccSource': 'SapEccSource', 'SapHanaSource': 'SapHanaSource', 'SapOpenHubSource': 'SapOpenHubSource', 'SapTableSource': 'SapTableSource', 'ServiceNowSource': 'ServiceNowSource', 'ShopifySource': 'ShopifySource', 'SparkSource': 'SparkSource', 'SqlDWSource': 'SqlDWSource', 'SqlMISource': 'SqlMISource', 'SqlServerSource': 'SqlServerSource', 'SqlSource': 'SqlSource', 'SquareSource': 'SquareSource', 'SybaseSource': 'SybaseSource', 'TeradataSource': 'TeradataSource', 'VerticaSource': 'VerticaSource', 'XeroSource': 'XeroSource', 'ZohoSource': 'ZohoSource'} - } - - def __init__( - self, - **kwargs - ): - super(TabularSource, self).__init__(**kwargs) - self.type = 'TabularSource' # type: str - self.query_timeout = kwargs.get('query_timeout', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class AmazonMWSSource(TabularSource): - """A copy activity Amazon Marketplace Web Service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AmazonMWSSource, self).__init__(**kwargs) - self.type = 'AmazonMWSSource' # type: str - self.query = kwargs.get('query', None) - - -class AmazonRedshiftLinkedService(LinkedService): - """Linked service for Amazon Redshift. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param server: Required. The name of the Amazon Redshift server. Type: string (or Expression - with resultType string). - :type server: any - :param username: The username of the Amazon Redshift source. Type: string (or Expression with - resultType string). - :type username: any - :param password: The password of the Amazon Redshift source. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param database: Required. The database name of the Amazon Redshift source. Type: string (or - Expression with resultType string). - :type database: any - :param port: The TCP port number that the Amazon Redshift server uses to listen for client - connections. The default value is 5439. Type: integer (or Expression with resultType integer). - :type port: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'server': {'required': True}, - 'database': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AmazonRedshiftLinkedService, self).__init__(**kwargs) - self.type = 'AmazonRedshift' # type: str - self.server = kwargs['server'] - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.database = kwargs['database'] - self.port = kwargs.get('port', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AmazonRedshiftSource(TabularSource): - """A copy activity source for Amazon Redshift Source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - :param redshift_unload_settings: The Amazon S3 settings needed for the interim Amazon S3 when - copying from Amazon Redshift with unload. With this, data from Amazon Redshift source will be - unloaded into S3 first and then copied into the targeted sink from the interim S3. - :type redshift_unload_settings: - ~azure.synapse.artifacts.v2020_12_01.models.RedshiftUnloadSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'redshift_unload_settings': {'key': 'redshiftUnloadSettings', 'type': 'RedshiftUnloadSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(AmazonRedshiftSource, self).__init__(**kwargs) - self.type = 'AmazonRedshiftSource' # type: str - self.query = kwargs.get('query', None) - self.redshift_unload_settings = kwargs.get('redshift_unload_settings', None) - - -class AmazonRedshiftTableDataset(Dataset): - """The Amazon Redshift table dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The Amazon Redshift table name. Type: string (or Expression with resultType - string). - :type table: any - :param schema_type_properties_schema: The Amazon Redshift schema name. Type: string (or - Expression with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AmazonRedshiftTableDataset, self).__init__(**kwargs) - self.type = 'AmazonRedshiftTable' # type: str - self.table_name = kwargs.get('table_name', None) - self.table = kwargs.get('table', None) - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - - -class AmazonS3Dataset(Dataset): - """A single Amazon Simple Storage Service (S3) object or a set of S3 objects. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param bucket_name: Required. The name of the Amazon S3 bucket. Type: string (or Expression - with resultType string). - :type bucket_name: any - :param key: The key of the Amazon S3 object. Type: string (or Expression with resultType - string). - :type key: any - :param prefix: The prefix filter for the S3 object name. Type: string (or Expression with - resultType string). - :type prefix: any - :param version: The version for the S3 object. Type: string (or Expression with resultType - string). - :type version: any - :param modified_datetime_start: The start of S3 object's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of S3 object's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_end: any - :param format: The format of files. - :type format: ~azure.synapse.artifacts.v2020_12_01.models.DatasetStorageFormat - :param compression: The data compression method used for the Amazon S3 object. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'bucket_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'bucket_name': {'key': 'typeProperties.bucketName', 'type': 'object'}, - 'key': {'key': 'typeProperties.key', 'type': 'object'}, - 'prefix': {'key': 'typeProperties.prefix', 'type': 'object'}, - 'version': {'key': 'typeProperties.version', 'type': 'object'}, - 'modified_datetime_start': {'key': 'typeProperties.modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'typeProperties.modifiedDatetimeEnd', 'type': 'object'}, - 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - } - - def __init__( - self, - **kwargs - ): - super(AmazonS3Dataset, self).__init__(**kwargs) - self.type = 'AmazonS3Object' # type: str - self.bucket_name = kwargs['bucket_name'] - self.key = kwargs.get('key', None) - self.prefix = kwargs.get('prefix', None) - self.version = kwargs.get('version', None) - self.modified_datetime_start = kwargs.get('modified_datetime_start', None) - self.modified_datetime_end = kwargs.get('modified_datetime_end', None) - self.format = kwargs.get('format', None) - self.compression = kwargs.get('compression', None) - - -class AmazonS3LinkedService(LinkedService): - """Linked service for Amazon S3. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param authentication_type: The authentication type of S3. Allowed value: AccessKey (default) - or TemporarySecurityCredentials. Type: string (or Expression with resultType string). - :type authentication_type: any - :param access_key_id: The access key identifier of the Amazon S3 Identity and Access Management - (IAM) user. Type: string (or Expression with resultType string). - :type access_key_id: any - :param secret_access_key: The secret access key of the Amazon S3 Identity and Access Management - (IAM) user. - :type secret_access_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param service_url: This value specifies the endpoint to access with the S3 Connector. This is - an optional property; change it only if you want to try a different service endpoint or want to - switch between https and http. Type: string (or Expression with resultType string). - :type service_url: any - :param session_token: The session token for the S3 temporary security credential. - :type session_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, - 'access_key_id': {'key': 'typeProperties.accessKeyId', 'type': 'object'}, - 'secret_access_key': {'key': 'typeProperties.secretAccessKey', 'type': 'SecretBase'}, - 'service_url': {'key': 'typeProperties.serviceUrl', 'type': 'object'}, - 'session_token': {'key': 'typeProperties.sessionToken', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AmazonS3LinkedService, self).__init__(**kwargs) - self.type = 'AmazonS3' # type: str - self.authentication_type = kwargs.get('authentication_type', None) - self.access_key_id = kwargs.get('access_key_id', None) - self.secret_access_key = kwargs.get('secret_access_key', None) - self.service_url = kwargs.get('service_url', None) - self.session_token = kwargs.get('session_token', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class DatasetLocation(msrest.serialization.Model): - """Dataset location. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonS3Location, AzureBlobFSLocation, AzureBlobStorageLocation, AzureDataLakeStoreLocation, AzureFileStorageLocation, FileServerLocation, FtpServerLocation, GoogleCloudStorageLocation, HdfsLocation, HttpServerLocation, SftpLocation. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - } - - _subtype_map = { - 'type': {'AmazonS3Location': 'AmazonS3Location', 'AzureBlobFSLocation': 'AzureBlobFSLocation', 'AzureBlobStorageLocation': 'AzureBlobStorageLocation', 'AzureDataLakeStoreLocation': 'AzureDataLakeStoreLocation', 'AzureFileStorageLocation': 'AzureFileStorageLocation', 'FileServerLocation': 'FileServerLocation', 'FtpServerLocation': 'FtpServerLocation', 'GoogleCloudStorageLocation': 'GoogleCloudStorageLocation', 'HdfsLocation': 'HdfsLocation', 'HttpServerLocation': 'HttpServerLocation', 'SftpLocation': 'SftpLocation'} - } - - def __init__( - self, - **kwargs - ): - super(DatasetLocation, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = 'DatasetLocation' # type: str - self.folder_path = kwargs.get('folder_path', None) - self.file_name = kwargs.get('file_name', None) - - -class AmazonS3Location(DatasetLocation): - """The location of amazon S3 dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - :param bucket_name: Specify the bucketName of amazon S3. Type: string (or Expression with - resultType string). - :type bucket_name: any - :param version: Specify the version of amazon S3. Type: string (or Expression with resultType - string). - :type version: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - 'bucket_name': {'key': 'bucketName', 'type': 'object'}, - 'version': {'key': 'version', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AmazonS3Location, self).__init__(**kwargs) - self.type = 'AmazonS3Location' # type: str - self.bucket_name = kwargs.get('bucket_name', None) - self.version = kwargs.get('version', None) - - -class StoreReadSettings(msrest.serialization.Model): - """Connector read setting. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonS3ReadSettings, AzureBlobFSReadSettings, AzureBlobStorageReadSettings, AzureDataLakeStoreReadSettings, AzureFileStorageReadSettings, FileServerReadSettings, FtpReadSettings, GoogleCloudStorageReadSettings, HdfsReadSettings, HttpReadSettings, SftpReadSettings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - } - - _subtype_map = { - 'type': {'AmazonS3ReadSettings': 'AmazonS3ReadSettings', 'AzureBlobFSReadSettings': 'AzureBlobFSReadSettings', 'AzureBlobStorageReadSettings': 'AzureBlobStorageReadSettings', 'AzureDataLakeStoreReadSettings': 'AzureDataLakeStoreReadSettings', 'AzureFileStorageReadSettings': 'AzureFileStorageReadSettings', 'FileServerReadSettings': 'FileServerReadSettings', 'FtpReadSettings': 'FtpReadSettings', 'GoogleCloudStorageReadSettings': 'GoogleCloudStorageReadSettings', 'HdfsReadSettings': 'HdfsReadSettings', 'HttpReadSettings': 'HttpReadSettings', 'SftpReadSettings': 'SftpReadSettings'} - } - - def __init__( - self, - **kwargs - ): - super(StoreReadSettings, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = 'StoreReadSettings' # type: str - self.max_concurrent_connections = kwargs.get('max_concurrent_connections', None) - - -class AmazonS3ReadSettings(StoreReadSettings): - """Azure data lake store read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: AmazonS3 wildcardFolderPath. Type: string (or Expression with - resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: AmazonS3 wildcardFileName. Type: string (or Expression with - resultType string). - :type wildcard_file_name: any - :param prefix: The prefix filter for the S3 object name. Type: string (or Expression with - resultType string). - :type prefix: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'prefix': {'key': 'prefix', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - 'modified_datetime_start': {'key': 'modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'modifiedDatetimeEnd', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AmazonS3ReadSettings, self).__init__(**kwargs) - self.type = 'AmazonS3ReadSettings' # type: str - self.recursive = kwargs.get('recursive', None) - self.wildcard_folder_path = kwargs.get('wildcard_folder_path', None) - self.wildcard_file_name = kwargs.get('wildcard_file_name', None) - self.prefix = kwargs.get('prefix', None) - self.file_list_path = kwargs.get('file_list_path', None) - self.enable_partition_discovery = kwargs.get('enable_partition_discovery', None) - self.partition_root_path = kwargs.get('partition_root_path', None) - self.delete_files_after_completion = kwargs.get('delete_files_after_completion', None) - self.modified_datetime_start = kwargs.get('modified_datetime_start', None) - self.modified_datetime_end = kwargs.get('modified_datetime_end', None) - - -class ControlActivity(Activity): - """Base class for all control activities like IfCondition, ForEach , Until. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AppendVariableActivity, ExecutePipelineActivity, FilterActivity, ForEachActivity, IfConditionActivity, SetVariableActivity, SwitchActivity, UntilActivity, ValidationActivity, WaitActivity, WebHookActivity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - } - - _subtype_map = { - 'type': {'AppendVariable': 'AppendVariableActivity', 'ExecutePipeline': 'ExecutePipelineActivity', 'Filter': 'FilterActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'SetVariable': 'SetVariableActivity', 'Switch': 'SwitchActivity', 'Until': 'UntilActivity', 'Validation': 'ValidationActivity', 'Wait': 'WaitActivity', 'WebHook': 'WebHookActivity'} - } - - def __init__( - self, - **kwargs - ): - super(ControlActivity, self).__init__(**kwargs) - self.type = 'Container' # type: str - - -class AppendVariableActivity(ControlActivity): - """Append value for a Variable of type Array. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param variable_name: Name of the variable whose value needs to be appended to. - :type variable_name: str - :param value: Value to be appended. Could be a static value or Expression. - :type value: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'variable_name': {'key': 'typeProperties.variableName', 'type': 'str'}, - 'value': {'key': 'typeProperties.value', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AppendVariableActivity, self).__init__(**kwargs) - self.type = 'AppendVariable' # type: str - self.variable_name = kwargs.get('variable_name', None) - self.value = kwargs.get('value', None) - - -class ArtifactRenameRequest(msrest.serialization.Model): - """Request body structure for rename artifact. - - :param new_name: New name of the artifact. - :type new_name: str - """ - - _validation = { - 'new_name': {'max_length': 260, 'min_length': 1, 'pattern': r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'}, - } - - _attribute_map = { - 'new_name': {'key': 'newName', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(ArtifactRenameRequest, self).__init__(**kwargs) - self.new_name = kwargs.get('new_name', None) - - -class AutoPauseProperties(msrest.serialization.Model): - """Auto-pausing properties of a Big Data pool powered by Apache Spark. - - :param delay_in_minutes: Number of minutes of idle time before the Big Data pool is - automatically paused. - :type delay_in_minutes: int - :param enabled: Whether auto-pausing is enabled for the Big Data pool. - :type enabled: bool - """ - - _attribute_map = { - 'delay_in_minutes': {'key': 'delayInMinutes', 'type': 'int'}, - 'enabled': {'key': 'enabled', 'type': 'bool'}, - } - - def __init__( - self, - **kwargs - ): - super(AutoPauseProperties, self).__init__(**kwargs) - self.delay_in_minutes = kwargs.get('delay_in_minutes', None) - self.enabled = kwargs.get('enabled', None) - - -class AutoScaleProperties(msrest.serialization.Model): - """Auto-scaling properties of a Big Data pool powered by Apache Spark. - - :param min_node_count: The minimum number of nodes the Big Data pool can support. - :type min_node_count: int - :param enabled: Whether automatic scaling is enabled for the Big Data pool. - :type enabled: bool - :param max_node_count: The maximum number of nodes the Big Data pool can support. - :type max_node_count: int - """ - - _attribute_map = { - 'min_node_count': {'key': 'minNodeCount', 'type': 'int'}, - 'enabled': {'key': 'enabled', 'type': 'bool'}, - 'max_node_count': {'key': 'maxNodeCount', 'type': 'int'}, - } - - def __init__( - self, - **kwargs - ): - super(AutoScaleProperties, self).__init__(**kwargs) - self.min_node_count = kwargs.get('min_node_count', None) - self.enabled = kwargs.get('enabled', None) - self.max_node_count = kwargs.get('max_node_count', None) - - -class AvroDataset(Dataset): - """Avro dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param location: The location of the avro storage. - :type location: ~azure.synapse.artifacts.v2020_12_01.models.DatasetLocation - :param avro_compression_codec: A string from AvroCompressionCodecEnum or an expression. - :type avro_compression_codec: any - :param avro_compression_level: - :type avro_compression_level: int - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'avro_compression_level': {'maximum': 9, 'minimum': 1}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'location': {'key': 'typeProperties.location', 'type': 'DatasetLocation'}, - 'avro_compression_codec': {'key': 'typeProperties.avroCompressionCodec', 'type': 'object'}, - 'avro_compression_level': {'key': 'typeProperties.avroCompressionLevel', 'type': 'int'}, - } - - def __init__( - self, - **kwargs - ): - super(AvroDataset, self).__init__(**kwargs) - self.type = 'Avro' # type: str - self.location = kwargs.get('location', None) - self.avro_compression_codec = kwargs.get('avro_compression_codec', None) - self.avro_compression_level = kwargs.get('avro_compression_level', None) - - -class DatasetStorageFormat(msrest.serialization.Model): - """The format definition of a storage. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AvroFormat, JsonFormat, OrcFormat, ParquetFormat, TextFormat. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage format.Constant filled by server. - :type type: str - :param serializer: Serializer. Type: string (or Expression with resultType string). - :type serializer: any - :param deserializer: Deserializer. Type: string (or Expression with resultType string). - :type deserializer: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'serializer': {'key': 'serializer', 'type': 'object'}, - 'deserializer': {'key': 'deserializer', 'type': 'object'}, - } - - _subtype_map = { - 'type': {'AvroFormat': 'AvroFormat', 'JsonFormat': 'JsonFormat', 'OrcFormat': 'OrcFormat', 'ParquetFormat': 'ParquetFormat', 'TextFormat': 'TextFormat'} - } - - def __init__( - self, - **kwargs - ): - super(DatasetStorageFormat, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = 'DatasetStorageFormat' # type: str - self.serializer = kwargs.get('serializer', None) - self.deserializer = kwargs.get('deserializer', None) - - -class AvroFormat(DatasetStorageFormat): - """The data stored in Avro format. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage format.Constant filled by server. - :type type: str - :param serializer: Serializer. Type: string (or Expression with resultType string). - :type serializer: any - :param deserializer: Deserializer. Type: string (or Expression with resultType string). - :type deserializer: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'serializer': {'key': 'serializer', 'type': 'object'}, - 'deserializer': {'key': 'deserializer', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AvroFormat, self).__init__(**kwargs) - self.type = 'AvroFormat' # type: str - - -class CopySink(msrest.serialization.Model): - """A copy activity sink. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AvroSink, AzureBlobFSSink, AzureDataExplorerSink, AzureDataLakeStoreSink, AzureDatabricksDeltaLakeSink, AzureMySqlSink, AzurePostgreSqlSink, AzureQueueSink, AzureSearchIndexSink, AzureSqlSink, AzureTableSink, BinarySink, BlobSink, CommonDataServiceForAppsSink, CosmosDbMongoDbApiSink, CosmosDbSqlApiSink, DelimitedTextSink, DocumentDbCollectionSink, DynamicsCrmSink, DynamicsSink, FileSystemSink, InformixSink, JsonSink, MicrosoftAccessSink, OdbcSink, OracleSink, OrcSink, ParquetSink, RestSink, SalesforceServiceCloudSink, SalesforceSink, SapCloudForCustomerSink, SnowflakeSink, SqlDWSink, SqlMISink, SqlServerSink, SqlSink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - } - - _subtype_map = { - 'type': {'AvroSink': 'AvroSink', 'AzureBlobFSSink': 'AzureBlobFSSink', 'AzureDataExplorerSink': 'AzureDataExplorerSink', 'AzureDataLakeStoreSink': 'AzureDataLakeStoreSink', 'AzureDatabricksDeltaLakeSink': 'AzureDatabricksDeltaLakeSink', 'AzureMySqlSink': 'AzureMySqlSink', 'AzurePostgreSqlSink': 'AzurePostgreSqlSink', 'AzureQueueSink': 'AzureQueueSink', 'AzureSearchIndexSink': 'AzureSearchIndexSink', 'AzureSqlSink': 'AzureSqlSink', 'AzureTableSink': 'AzureTableSink', 'BinarySink': 'BinarySink', 'BlobSink': 'BlobSink', 'CommonDataServiceForAppsSink': 'CommonDataServiceForAppsSink', 'CosmosDbMongoDbApiSink': 'CosmosDbMongoDbApiSink', 'CosmosDbSqlApiSink': 'CosmosDbSqlApiSink', 'DelimitedTextSink': 'DelimitedTextSink', 'DocumentDbCollectionSink': 'DocumentDbCollectionSink', 'DynamicsCrmSink': 'DynamicsCrmSink', 'DynamicsSink': 'DynamicsSink', 'FileSystemSink': 'FileSystemSink', 'InformixSink': 'InformixSink', 'JsonSink': 'JsonSink', 'MicrosoftAccessSink': 'MicrosoftAccessSink', 'OdbcSink': 'OdbcSink', 'OracleSink': 'OracleSink', 'OrcSink': 'OrcSink', 'ParquetSink': 'ParquetSink', 'RestSink': 'RestSink', 'SalesforceServiceCloudSink': 'SalesforceServiceCloudSink', 'SalesforceSink': 'SalesforceSink', 'SapCloudForCustomerSink': 'SapCloudForCustomerSink', 'SnowflakeSink': 'SnowflakeSink', 'SqlDWSink': 'SqlDWSink', 'SqlMISink': 'SqlMISink', 'SqlServerSink': 'SqlServerSink', 'SqlSink': 'SqlSink'} - } - - def __init__( - self, - **kwargs - ): - super(CopySink, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = 'CopySink' # type: str - self.write_batch_size = kwargs.get('write_batch_size', None) - self.write_batch_timeout = kwargs.get('write_batch_timeout', None) - self.sink_retry_count = kwargs.get('sink_retry_count', None) - self.sink_retry_wait = kwargs.get('sink_retry_wait', None) - self.max_concurrent_connections = kwargs.get('max_concurrent_connections', None) - - -class AvroSink(CopySink): - """A copy activity Avro sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Avro store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreWriteSettings - :param format_settings: Avro format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.AvroWriteSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreWriteSettings'}, - 'format_settings': {'key': 'formatSettings', 'type': 'AvroWriteSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(AvroSink, self).__init__(**kwargs) - self.type = 'AvroSink' # type: str - self.store_settings = kwargs.get('store_settings', None) - self.format_settings = kwargs.get('format_settings', None) - - -class AvroSource(CopySource): - """A copy activity Avro source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Avro store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AvroSource, self).__init__(**kwargs) - self.type = 'AvroSource' # type: str - self.store_settings = kwargs.get('store_settings', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class FormatWriteSettings(msrest.serialization.Model): - """Format write settings. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AvroWriteSettings, DelimitedTextWriteSettings, JsonWriteSettings, OrcWriteSettings, ParquetWriteSettings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - _subtype_map = { - 'type': {'AvroWriteSettings': 'AvroWriteSettings', 'DelimitedTextWriteSettings': 'DelimitedTextWriteSettings', 'JsonWriteSettings': 'JsonWriteSettings', 'OrcWriteSettings': 'OrcWriteSettings', 'ParquetWriteSettings': 'ParquetWriteSettings'} - } - - def __init__( - self, - **kwargs - ): - super(FormatWriteSettings, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = 'FormatWriteSettings' # type: str - - -class AvroWriteSettings(FormatWriteSettings): - """Avro write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param record_name: Top level record name in write result, which is required in AVRO spec. - :type record_name: str - :param record_namespace: Record namespace in the write result. - :type record_namespace: str - :param max_rows_per_file: Limit the written file's row count to be smaller than or equal to the - specified count. Type: integer (or Expression with resultType integer). - :type max_rows_per_file: any - :param file_name_prefix: Specifies the file name pattern - :code:``_:code:``.:code:`` when copy from non-file - based store without partitionOptions. Type: string (or Expression with resultType string). - :type file_name_prefix: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'record_name': {'key': 'recordName', 'type': 'str'}, - 'record_namespace': {'key': 'recordNamespace', 'type': 'str'}, - 'max_rows_per_file': {'key': 'maxRowsPerFile', 'type': 'object'}, - 'file_name_prefix': {'key': 'fileNamePrefix', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AvroWriteSettings, self).__init__(**kwargs) - self.type = 'AvroWriteSettings' # type: str - self.record_name = kwargs.get('record_name', None) - self.record_namespace = kwargs.get('record_namespace', None) - self.max_rows_per_file = kwargs.get('max_rows_per_file', None) - self.file_name_prefix = kwargs.get('file_name_prefix', None) - - -class AzureBatchLinkedService(LinkedService): - """Azure Batch linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param account_name: Required. The Azure Batch account name. Type: string (or Expression with - resultType string). - :type account_name: any - :param access_key: The Azure Batch account access key. - :type access_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param batch_uri: Required. The Azure Batch URI. Type: string (or Expression with resultType - string). - :type batch_uri: any - :param pool_name: Required. The Azure Batch pool name. Type: string (or Expression with - resultType string). - :type pool_name: any - :param linked_service_name: Required. The Azure Storage linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'account_name': {'required': True}, - 'batch_uri': {'required': True}, - 'pool_name': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'account_name': {'key': 'typeProperties.accountName', 'type': 'object'}, - 'access_key': {'key': 'typeProperties.accessKey', 'type': 'SecretBase'}, - 'batch_uri': {'key': 'typeProperties.batchUri', 'type': 'object'}, - 'pool_name': {'key': 'typeProperties.poolName', 'type': 'object'}, - 'linked_service_name': {'key': 'typeProperties.linkedServiceName', 'type': 'LinkedServiceReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureBatchLinkedService, self).__init__(**kwargs) - self.type = 'AzureBatch' # type: str - self.account_name = kwargs['account_name'] - self.access_key = kwargs.get('access_key', None) - self.batch_uri = kwargs['batch_uri'] - self.pool_name = kwargs['pool_name'] - self.linked_service_name = kwargs['linked_service_name'] - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AzureBlobDataset(Dataset): - """The Azure Blob storage. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param folder_path: The path of the Azure Blob storage. Type: string (or Expression with - resultType string). - :type folder_path: any - :param table_root_location: The root of blob path. Type: string (or Expression with resultType - string). - :type table_root_location: any - :param file_name: The name of the Azure Blob. Type: string (or Expression with resultType - string). - :type file_name: any - :param modified_datetime_start: The start of Azure Blob's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of Azure Blob's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_end: any - :param format: The format of the Azure Blob storage. - :type format: ~azure.synapse.artifacts.v2020_12_01.models.DatasetStorageFormat - :param compression: The data compression method used for the blob storage. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, - 'table_root_location': {'key': 'typeProperties.tableRootLocation', 'type': 'object'}, - 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, - 'modified_datetime_start': {'key': 'typeProperties.modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'typeProperties.modifiedDatetimeEnd', 'type': 'object'}, - 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureBlobDataset, self).__init__(**kwargs) - self.type = 'AzureBlob' # type: str - self.folder_path = kwargs.get('folder_path', None) - self.table_root_location = kwargs.get('table_root_location', None) - self.file_name = kwargs.get('file_name', None) - self.modified_datetime_start = kwargs.get('modified_datetime_start', None) - self.modified_datetime_end = kwargs.get('modified_datetime_end', None) - self.format = kwargs.get('format', None) - self.compression = kwargs.get('compression', None) - - -class AzureBlobFSDataset(Dataset): - """The Azure Data Lake Storage Gen2 storage. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param folder_path: The path of the Azure Data Lake Storage Gen2 storage. Type: string (or - Expression with resultType string). - :type folder_path: any - :param file_name: The name of the Azure Data Lake Storage Gen2. Type: string (or Expression - with resultType string). - :type file_name: any - :param format: The format of the Azure Data Lake Storage Gen2 storage. - :type format: ~azure.synapse.artifacts.v2020_12_01.models.DatasetStorageFormat - :param compression: The data compression method used for the blob storage. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, - 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, - 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureBlobFSDataset, self).__init__(**kwargs) - self.type = 'AzureBlobFSFile' # type: str - self.folder_path = kwargs.get('folder_path', None) - self.file_name = kwargs.get('file_name', None) - self.format = kwargs.get('format', None) - self.compression = kwargs.get('compression', None) - - -class AzureBlobFSLinkedService(LinkedService): - """Azure Data Lake Storage Gen2 linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param url: Required. Endpoint for the Azure Data Lake Storage Gen2 service. Type: string (or - Expression with resultType string). - :type url: any - :param account_key: Account key for the Azure Data Lake Storage Gen2 service. Type: string (or - Expression with resultType string). - :type account_key: any - :param service_principal_id: The ID of the application used to authenticate against the Azure - Data Lake Storage Gen2 account. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: The Key of the application used to authenticate against the Azure - Data Lake Storage Gen2 account. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: The name or ID of the tenant to which the service principal belongs. Type: - string (or Expression with resultType string). - :type tenant: any - :param azure_cloud_type: Indicates the azure cloud type of the service principle auth. Allowed - values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data - factory regions’ cloud type. Type: string (or Expression with resultType string). - :type azure_cloud_type: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'account_key': {'key': 'typeProperties.accountKey', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'azure_cloud_type': {'key': 'typeProperties.azureCloudType', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureBlobFSLinkedService, self).__init__(**kwargs) - self.type = 'AzureBlobFS' # type: str - self.url = kwargs['url'] - self.account_key = kwargs.get('account_key', None) - self.service_principal_id = kwargs.get('service_principal_id', None) - self.service_principal_key = kwargs.get('service_principal_key', None) - self.tenant = kwargs.get('tenant', None) - self.azure_cloud_type = kwargs.get('azure_cloud_type', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AzureBlobFSLocation(DatasetLocation): - """The location of azure blobFS dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - :param file_system: Specify the fileSystem of azure blobFS. Type: string (or Expression with - resultType string). - :type file_system: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - 'file_system': {'key': 'fileSystem', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureBlobFSLocation, self).__init__(**kwargs) - self.type = 'AzureBlobFSLocation' # type: str - self.file_system = kwargs.get('file_system', None) - - -class AzureBlobFSReadSettings(StoreReadSettings): - """Azure blobFS read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: Azure blobFS wildcardFolderPath. Type: string (or Expression with - resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: Azure blobFS wildcardFileName. Type: string (or Expression with - resultType string). - :type wildcard_file_name: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - 'modified_datetime_start': {'key': 'modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'modifiedDatetimeEnd', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureBlobFSReadSettings, self).__init__(**kwargs) - self.type = 'AzureBlobFSReadSettings' # type: str - self.recursive = kwargs.get('recursive', None) - self.wildcard_folder_path = kwargs.get('wildcard_folder_path', None) - self.wildcard_file_name = kwargs.get('wildcard_file_name', None) - self.file_list_path = kwargs.get('file_list_path', None) - self.enable_partition_discovery = kwargs.get('enable_partition_discovery', None) - self.partition_root_path = kwargs.get('partition_root_path', None) - self.delete_files_after_completion = kwargs.get('delete_files_after_completion', None) - self.modified_datetime_start = kwargs.get('modified_datetime_start', None) - self.modified_datetime_end = kwargs.get('modified_datetime_end', None) - - -class AzureBlobFSSink(CopySink): - """A copy activity Azure Data Lake Storage Gen2 sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureBlobFSSink, self).__init__(**kwargs) - self.type = 'AzureBlobFSSink' # type: str - self.copy_behavior = kwargs.get('copy_behavior', None) - - -class AzureBlobFSSource(CopySource): - """A copy activity Azure BlobFS source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param treat_empty_as_null: Treat empty as null. Type: boolean (or Expression with resultType - boolean). - :type treat_empty_as_null: any - :param skip_header_line_count: Number of header lines to skip from each blob. Type: integer (or - Expression with resultType integer). - :type skip_header_line_count: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'treat_empty_as_null': {'key': 'treatEmptyAsNull', 'type': 'object'}, - 'skip_header_line_count': {'key': 'skipHeaderLineCount', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureBlobFSSource, self).__init__(**kwargs) - self.type = 'AzureBlobFSSource' # type: str - self.treat_empty_as_null = kwargs.get('treat_empty_as_null', None) - self.skip_header_line_count = kwargs.get('skip_header_line_count', None) - self.recursive = kwargs.get('recursive', None) - - -class StoreWriteSettings(msrest.serialization.Model): - """Connector write settings. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AzureBlobFSWriteSettings, AzureBlobStorageWriteSettings, AzureDataLakeStoreWriteSettings, AzureFileStorageWriteSettings, FileServerWriteSettings, SftpWriteSettings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - } - - _subtype_map = { - 'type': {'AzureBlobFSWriteSettings': 'AzureBlobFSWriteSettings', 'AzureBlobStorageWriteSettings': 'AzureBlobStorageWriteSettings', 'AzureDataLakeStoreWriteSettings': 'AzureDataLakeStoreWriteSettings', 'AzureFileStorageWriteSettings': 'AzureFileStorageWriteSettings', 'FileServerWriteSettings': 'FileServerWriteSettings', 'SftpWriteSettings': 'SftpWriteSettings'} - } - - def __init__( - self, - **kwargs - ): - super(StoreWriteSettings, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = 'StoreWriteSettings' # type: str - self.max_concurrent_connections = kwargs.get('max_concurrent_connections', None) - self.copy_behavior = kwargs.get('copy_behavior', None) - - -class AzureBlobFSWriteSettings(StoreWriteSettings): - """Azure blobFS write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - :param block_size_in_mb: Indicates the block size(MB) when writing data to blob. Type: integer - (or Expression with resultType integer). - :type block_size_in_mb: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - 'block_size_in_mb': {'key': 'blockSizeInMB', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureBlobFSWriteSettings, self).__init__(**kwargs) - self.type = 'AzureBlobFSWriteSettings' # type: str - self.block_size_in_mb = kwargs.get('block_size_in_mb', None) - - -class AzureBlobStorageLinkedService(LinkedService): - """The azure blob storage linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: The connection string. It is mutually exclusive with sasUri, - serviceEndpoint property. Type: string, SecureString or AzureKeyVaultSecretReference. - :type connection_string: any - :param account_key: The Azure key vault secret reference of accountKey in connection string. - :type account_key: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param sas_uri: SAS URI of the Azure Blob Storage resource. It is mutually exclusive with - connectionString, serviceEndpoint property. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type sas_uri: any - :param sas_token: The Azure key vault secret reference of sasToken in sas uri. - :type sas_token: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param service_endpoint: Blob service endpoint of the Azure Blob Storage resource. It is - mutually exclusive with connectionString, sasUri property. - :type service_endpoint: str - :param service_principal_id: The ID of the service principal used to authenticate against Azure - SQL Data Warehouse. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: The key of the service principal used to authenticate against - Azure SQL Data Warehouse. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: The name or ID of the tenant to which the service principal belongs. Type: - string (or Expression with resultType string). - :type tenant: any - :param azure_cloud_type: Indicates the azure cloud type of the service principle auth. Allowed - values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data - factory regions’ cloud type. Type: string (or Expression with resultType string). - :type azure_cloud_type: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'account_key': {'key': 'typeProperties.accountKey', 'type': 'AzureKeyVaultSecretReference'}, - 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'object'}, - 'sas_token': {'key': 'typeProperties.sasToken', 'type': 'AzureKeyVaultSecretReference'}, - 'service_endpoint': {'key': 'typeProperties.serviceEndpoint', 'type': 'str'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'azure_cloud_type': {'key': 'typeProperties.azureCloudType', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureBlobStorageLinkedService, self).__init__(**kwargs) - self.type = 'AzureBlobStorage' # type: str - self.connection_string = kwargs.get('connection_string', None) - self.account_key = kwargs.get('account_key', None) - self.sas_uri = kwargs.get('sas_uri', None) - self.sas_token = kwargs.get('sas_token', None) - self.service_endpoint = kwargs.get('service_endpoint', None) - self.service_principal_id = kwargs.get('service_principal_id', None) - self.service_principal_key = kwargs.get('service_principal_key', None) - self.tenant = kwargs.get('tenant', None) - self.azure_cloud_type = kwargs.get('azure_cloud_type', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AzureBlobStorageLocation(DatasetLocation): - """The location of azure blob dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - :param container: Specify the container of azure blob. Type: string (or Expression with - resultType string). - :type container: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - 'container': {'key': 'container', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureBlobStorageLocation, self).__init__(**kwargs) - self.type = 'AzureBlobStorageLocation' # type: str - self.container = kwargs.get('container', None) - - -class AzureBlobStorageReadSettings(StoreReadSettings): - """Azure blob read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: Azure blob wildcardFolderPath. Type: string (or Expression with - resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: Azure blob wildcardFileName. Type: string (or Expression with - resultType string). - :type wildcard_file_name: any - :param prefix: The prefix filter for the Azure Blob name. Type: string (or Expression with - resultType string). - :type prefix: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'prefix': {'key': 'prefix', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - 'modified_datetime_start': {'key': 'modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'modifiedDatetimeEnd', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureBlobStorageReadSettings, self).__init__(**kwargs) - self.type = 'AzureBlobStorageReadSettings' # type: str - self.recursive = kwargs.get('recursive', None) - self.wildcard_folder_path = kwargs.get('wildcard_folder_path', None) - self.wildcard_file_name = kwargs.get('wildcard_file_name', None) - self.prefix = kwargs.get('prefix', None) - self.file_list_path = kwargs.get('file_list_path', None) - self.enable_partition_discovery = kwargs.get('enable_partition_discovery', None) - self.partition_root_path = kwargs.get('partition_root_path', None) - self.delete_files_after_completion = kwargs.get('delete_files_after_completion', None) - self.modified_datetime_start = kwargs.get('modified_datetime_start', None) - self.modified_datetime_end = kwargs.get('modified_datetime_end', None) - - -class AzureBlobStorageWriteSettings(StoreWriteSettings): - """Azure blob write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - :param block_size_in_mb: Indicates the block size(MB) when writing data to blob. Type: integer - (or Expression with resultType integer). - :type block_size_in_mb: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - 'block_size_in_mb': {'key': 'blockSizeInMB', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureBlobStorageWriteSettings, self).__init__(**kwargs) - self.type = 'AzureBlobStorageWriteSettings' # type: str - self.block_size_in_mb = kwargs.get('block_size_in_mb', None) - - -class AzureDatabricksDeltaLakeDataset(Dataset): - """Azure Databricks Delta Lake dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table: The name of delta table. Type: string (or Expression with resultType string). - :type table: any - :param database: The database name of delta table. Type: string (or Expression with resultType - string). - :type database: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDatabricksDeltaLakeDataset, self).__init__(**kwargs) - self.type = 'AzureDatabricksDeltaLakeDataset' # type: str - self.table = kwargs.get('table', None) - self.database = kwargs.get('database', None) - - -class ExportSettings(msrest.serialization.Model): - """Export command settings. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AzureDatabricksDeltaLakeExportCommand, SnowflakeExportCopyCommand. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The export setting type.Constant filled by server. - :type type: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - _subtype_map = { - 'type': {'AzureDatabricksDeltaLakeExportCommand': 'AzureDatabricksDeltaLakeExportCommand', 'SnowflakeExportCopyCommand': 'SnowflakeExportCopyCommand'} - } - - def __init__( - self, - **kwargs - ): - super(ExportSettings, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = 'ExportSettings' # type: str - - -class AzureDatabricksDeltaLakeExportCommand(ExportSettings): - """Azure Databricks Delta Lake export command settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The export setting type.Constant filled by server. - :type type: str - :param date_format: Specify the date format for the csv in Azure Databricks Delta Lake Copy. - Type: string (or Expression with resultType string). - :type date_format: any - :param timestamp_format: Specify the timestamp format for the csv in Azure Databricks Delta - Lake Copy. Type: string (or Expression with resultType string). - :type timestamp_format: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'date_format': {'key': 'dateFormat', 'type': 'object'}, - 'timestamp_format': {'key': 'timestampFormat', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDatabricksDeltaLakeExportCommand, self).__init__(**kwargs) - self.type = 'AzureDatabricksDeltaLakeExportCommand' # type: str - self.date_format = kwargs.get('date_format', None) - self.timestamp_format = kwargs.get('timestamp_format', None) - - -class ImportSettings(msrest.serialization.Model): - """Import command settings. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AzureDatabricksDeltaLakeImportCommand, SnowflakeImportCopyCommand. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The import setting type.Constant filled by server. - :type type: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - _subtype_map = { - 'type': {'AzureDatabricksDeltaLakeImportCommand': 'AzureDatabricksDeltaLakeImportCommand', 'SnowflakeImportCopyCommand': 'SnowflakeImportCopyCommand'} - } - - def __init__( - self, - **kwargs - ): - super(ImportSettings, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = 'ImportSettings' # type: str - - -class AzureDatabricksDeltaLakeImportCommand(ImportSettings): - """Azure Databricks Delta Lake import command settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The import setting type.Constant filled by server. - :type type: str - :param date_format: Specify the date format for csv in Azure Databricks Delta Lake Copy. Type: - string (or Expression with resultType string). - :type date_format: any - :param timestamp_format: Specify the timestamp format for csv in Azure Databricks Delta Lake - Copy. Type: string (or Expression with resultType string). - :type timestamp_format: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'date_format': {'key': 'dateFormat', 'type': 'object'}, - 'timestamp_format': {'key': 'timestampFormat', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDatabricksDeltaLakeImportCommand, self).__init__(**kwargs) - self.type = 'AzureDatabricksDeltaLakeImportCommand' # type: str - self.date_format = kwargs.get('date_format', None) - self.timestamp_format = kwargs.get('timestamp_format', None) - - -class AzureDatabricksDeltaLakeLinkedService(LinkedService): - """Azure Databricks Delta Lake linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param domain: Required. :code:``.azuredatabricks.net, domain name of your Databricks - deployment. Type: string (or Expression with resultType string). - :type domain: any - :param access_token: Required. Access token for databricks REST API. Refer to - https://docs.azuredatabricks.net/api/latest/authentication.html. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type access_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param cluster_id: The id of an existing interactive cluster that will be used for all runs of - this job. Type: string (or Expression with resultType string). - :type cluster_id: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'domain': {'required': True}, - 'access_token': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'domain': {'key': 'typeProperties.domain', 'type': 'object'}, - 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, - 'cluster_id': {'key': 'typeProperties.clusterId', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDatabricksDeltaLakeLinkedService, self).__init__(**kwargs) - self.type = 'AzureDatabricksDeltaLake' # type: str - self.domain = kwargs['domain'] - self.access_token = kwargs['access_token'] - self.cluster_id = kwargs.get('cluster_id', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AzureDatabricksDeltaLakeSink(CopySink): - """A copy activity Azure Databricks Delta Lake sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType - string). - :type pre_copy_script: any - :param import_settings: Azure Databricks Delta Lake import settings. - :type import_settings: - ~azure.synapse.artifacts.v2020_12_01.models.AzureDatabricksDeltaLakeImportCommand - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - 'import_settings': {'key': 'importSettings', 'type': 'AzureDatabricksDeltaLakeImportCommand'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDatabricksDeltaLakeSink, self).__init__(**kwargs) - self.type = 'AzureDatabricksDeltaLakeSink' # type: str - self.pre_copy_script = kwargs.get('pre_copy_script', None) - self.import_settings = kwargs.get('import_settings', None) - - -class AzureDatabricksDeltaLakeSource(CopySource): - """A copy activity Azure Databricks Delta Lake source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: Azure Databricks Delta Lake Sql query. Type: string (or Expression with - resultType string). - :type query: any - :param export_settings: Azure Databricks Delta Lake export settings. - :type export_settings: - ~azure.synapse.artifacts.v2020_12_01.models.AzureDatabricksDeltaLakeExportCommand - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'export_settings': {'key': 'exportSettings', 'type': 'AzureDatabricksDeltaLakeExportCommand'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDatabricksDeltaLakeSource, self).__init__(**kwargs) - self.type = 'AzureDatabricksDeltaLakeSource' # type: str - self.query = kwargs.get('query', None) - self.export_settings = kwargs.get('export_settings', None) - - -class AzureDatabricksLinkedService(LinkedService): - """Azure Databricks linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param domain: Required. :code:``.azuredatabricks.net, domain name of your Databricks - deployment. Type: string (or Expression with resultType string). - :type domain: any - :param access_token: Access token for databricks REST API. Refer to - https://docs.azuredatabricks.net/api/latest/authentication.html. Type: string (or Expression - with resultType string). - :type access_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param authentication: Required to specify MSI, if using Workspace resource id for databricks - REST API. Type: string (or Expression with resultType string). - :type authentication: any - :param workspace_resource_id: Workspace resource id for databricks REST API. Type: string (or - Expression with resultType string). - :type workspace_resource_id: any - :param existing_cluster_id: The id of an existing interactive cluster that will be used for all - runs of this activity. Type: string (or Expression with resultType string). - :type existing_cluster_id: any - :param instance_pool_id: The id of an existing instance pool that will be used for all runs of - this activity. Type: string (or Expression with resultType string). - :type instance_pool_id: any - :param new_cluster_version: If not using an existing interactive cluster, this specifies the - Spark version of a new job cluster or instance pool nodes created for each run of this - activity. Required if instancePoolId is specified. Type: string (or Expression with resultType - string). - :type new_cluster_version: any - :param new_cluster_num_of_worker: If not using an existing interactive cluster, this specifies - the number of worker nodes to use for the new job cluster or instance pool. For new job - clusters, this a string-formatted Int32, like '1' means numOfWorker is 1 or '1:10' means - auto-scale from 1 (min) to 10 (max). For instance pools, this is a string-formatted Int32, and - can only specify a fixed number of worker nodes, such as '2'. Required if newClusterVersion is - specified. Type: string (or Expression with resultType string). - :type new_cluster_num_of_worker: any - :param new_cluster_node_type: The node type of the new job cluster. This property is required - if newClusterVersion is specified and instancePoolId is not specified. If instancePoolId is - specified, this property is ignored. Type: string (or Expression with resultType string). - :type new_cluster_node_type: any - :param new_cluster_spark_conf: A set of optional, user-specified Spark configuration key-value - pairs. - :type new_cluster_spark_conf: dict[str, any] - :param new_cluster_spark_env_vars: A set of optional, user-specified Spark environment - variables key-value pairs. - :type new_cluster_spark_env_vars: dict[str, any] - :param new_cluster_custom_tags: Additional tags for cluster resources. This property is ignored - in instance pool configurations. - :type new_cluster_custom_tags: dict[str, any] - :param new_cluster_log_destination: Specify a location to deliver Spark driver, worker, and - event logs. Type: string (or Expression with resultType string). - :type new_cluster_log_destination: any - :param new_cluster_driver_node_type: The driver node type for the new job cluster. This - property is ignored in instance pool configurations. Type: string (or Expression with - resultType string). - :type new_cluster_driver_node_type: any - :param new_cluster_init_scripts: User-defined initialization scripts for the new cluster. Type: - array of strings (or Expression with resultType array of strings). - :type new_cluster_init_scripts: any - :param new_cluster_enable_elastic_disk: Enable the elastic disk on the new cluster. This - property is now ignored, and takes the default elastic disk behavior in Databricks (elastic - disks are always enabled). Type: boolean (or Expression with resultType boolean). - :type new_cluster_enable_elastic_disk: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - :param policy_id: The policy id for limiting the ability to configure clusters based on a user - defined set of rules. Type: string (or Expression with resultType string). - :type policy_id: any - """ - - _validation = { - 'type': {'required': True}, - 'domain': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'domain': {'key': 'typeProperties.domain', 'type': 'object'}, - 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, - 'authentication': {'key': 'typeProperties.authentication', 'type': 'object'}, - 'workspace_resource_id': {'key': 'typeProperties.workspaceResourceId', 'type': 'object'}, - 'existing_cluster_id': {'key': 'typeProperties.existingClusterId', 'type': 'object'}, - 'instance_pool_id': {'key': 'typeProperties.instancePoolId', 'type': 'object'}, - 'new_cluster_version': {'key': 'typeProperties.newClusterVersion', 'type': 'object'}, - 'new_cluster_num_of_worker': {'key': 'typeProperties.newClusterNumOfWorker', 'type': 'object'}, - 'new_cluster_node_type': {'key': 'typeProperties.newClusterNodeType', 'type': 'object'}, - 'new_cluster_spark_conf': {'key': 'typeProperties.newClusterSparkConf', 'type': '{object}'}, - 'new_cluster_spark_env_vars': {'key': 'typeProperties.newClusterSparkEnvVars', 'type': '{object}'}, - 'new_cluster_custom_tags': {'key': 'typeProperties.newClusterCustomTags', 'type': '{object}'}, - 'new_cluster_log_destination': {'key': 'typeProperties.newClusterLogDestination', 'type': 'object'}, - 'new_cluster_driver_node_type': {'key': 'typeProperties.newClusterDriverNodeType', 'type': 'object'}, - 'new_cluster_init_scripts': {'key': 'typeProperties.newClusterInitScripts', 'type': 'object'}, - 'new_cluster_enable_elastic_disk': {'key': 'typeProperties.newClusterEnableElasticDisk', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - 'policy_id': {'key': 'typeProperties.policyId', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDatabricksLinkedService, self).__init__(**kwargs) - self.type = 'AzureDatabricks' # type: str - self.domain = kwargs['domain'] - self.access_token = kwargs.get('access_token', None) - self.authentication = kwargs.get('authentication', None) - self.workspace_resource_id = kwargs.get('workspace_resource_id', None) - self.existing_cluster_id = kwargs.get('existing_cluster_id', None) - self.instance_pool_id = kwargs.get('instance_pool_id', None) - self.new_cluster_version = kwargs.get('new_cluster_version', None) - self.new_cluster_num_of_worker = kwargs.get('new_cluster_num_of_worker', None) - self.new_cluster_node_type = kwargs.get('new_cluster_node_type', None) - self.new_cluster_spark_conf = kwargs.get('new_cluster_spark_conf', None) - self.new_cluster_spark_env_vars = kwargs.get('new_cluster_spark_env_vars', None) - self.new_cluster_custom_tags = kwargs.get('new_cluster_custom_tags', None) - self.new_cluster_log_destination = kwargs.get('new_cluster_log_destination', None) - self.new_cluster_driver_node_type = kwargs.get('new_cluster_driver_node_type', None) - self.new_cluster_init_scripts = kwargs.get('new_cluster_init_scripts', None) - self.new_cluster_enable_elastic_disk = kwargs.get('new_cluster_enable_elastic_disk', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - self.policy_id = kwargs.get('policy_id', None) - - -class ExecutionActivity(Activity): - """Base class for all execution activities. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AzureDataExplorerCommandActivity, AzureFunctionActivity, AzureMLBatchExecutionActivity, AzureMLExecutePipelineActivity, AzureMLUpdateResourceActivity, CopyActivity, CustomActivity, DataLakeAnalyticsUSQLActivity, DatabricksNotebookActivity, DatabricksSparkJarActivity, DatabricksSparkPythonActivity, DeleteActivity, ExecuteDataFlowActivity, ExecuteSSISPackageActivity, GetMetadataActivity, HDInsightHiveActivity, HDInsightMapReduceActivity, HDInsightPigActivity, HDInsightSparkActivity, HDInsightStreamingActivity, LookupActivity, SynapseSparkJobDefinitionActivity, SqlServerStoredProcedureActivity, SynapseNotebookActivity, WebActivity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - } - - _subtype_map = { - 'type': {'AzureDataExplorerCommand': 'AzureDataExplorerCommandActivity', 'AzureFunctionActivity': 'AzureFunctionActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'AzureMLExecutePipeline': 'AzureMLExecutePipelineActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'Copy': 'CopyActivity', 'Custom': 'CustomActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'DatabricksNotebook': 'DatabricksNotebookActivity', 'DatabricksSparkJar': 'DatabricksSparkJarActivity', 'DatabricksSparkPython': 'DatabricksSparkPythonActivity', 'Delete': 'DeleteActivity', 'ExecuteDataFlow': 'ExecuteDataFlowActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'GetMetadata': 'GetMetadataActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'Lookup': 'LookupActivity', 'SparkJob': 'SynapseSparkJobDefinitionActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'SynapseNotebook': 'SynapseNotebookActivity', 'WebActivity': 'WebActivity'} - } - - def __init__( - self, - **kwargs - ): - super(ExecutionActivity, self).__init__(**kwargs) - self.type = 'Execution' # type: str - self.linked_service_name = kwargs.get('linked_service_name', None) - self.policy = kwargs.get('policy', None) - - -class AzureDataExplorerCommandActivity(ExecutionActivity): - """Azure Data Explorer command activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param command: Required. A control command, according to the Azure Data Explorer command - syntax. Type: string (or Expression with resultType string). - :type command: any - :param command_timeout: Control command timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))..). - :type command_timeout: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'command': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'command': {'key': 'typeProperties.command', 'type': 'object'}, - 'command_timeout': {'key': 'typeProperties.commandTimeout', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDataExplorerCommandActivity, self).__init__(**kwargs) - self.type = 'AzureDataExplorerCommand' # type: str - self.command = kwargs['command'] - self.command_timeout = kwargs.get('command_timeout', None) - - -class AzureDataExplorerLinkedService(LinkedService): - """Azure Data Explorer (Kusto) linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param endpoint: Required. The endpoint of Azure Data Explorer (the engine's endpoint). URL - will be in the format https://:code:``.:code:``.kusto.windows.net. - Type: string (or Expression with resultType string). - :type endpoint: any - :param service_principal_id: Required. The ID of the service principal used to authenticate - against Azure Data Explorer. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: Required. The key of the service principal used to authenticate - against Kusto. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param database: Required. Database name for connection. Type: string (or Expression with - resultType string). - :type database: any - :param tenant: Required. The name or ID of the tenant to which the service principal belongs. - Type: string (or Expression with resultType string). - :type tenant: any - """ - - _validation = { - 'type': {'required': True}, - 'endpoint': {'required': True}, - 'service_principal_id': {'required': True}, - 'service_principal_key': {'required': True}, - 'database': {'required': True}, - 'tenant': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDataExplorerLinkedService, self).__init__(**kwargs) - self.type = 'AzureDataExplorer' # type: str - self.endpoint = kwargs['endpoint'] - self.service_principal_id = kwargs['service_principal_id'] - self.service_principal_key = kwargs['service_principal_key'] - self.database = kwargs['database'] - self.tenant = kwargs['tenant'] - - -class AzureDataExplorerSink(CopySink): - """A copy activity Azure Data Explorer sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param ingestion_mapping_name: A name of a pre-created csv mapping that was defined on the - target Kusto table. Type: string. - :type ingestion_mapping_name: any - :param ingestion_mapping_as_json: An explicit column mapping description provided in a json - format. Type: string. - :type ingestion_mapping_as_json: any - :param flush_immediately: If set to true, any aggregation will be skipped. Default is false. - Type: boolean. - :type flush_immediately: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'ingestion_mapping_name': {'key': 'ingestionMappingName', 'type': 'object'}, - 'ingestion_mapping_as_json': {'key': 'ingestionMappingAsJson', 'type': 'object'}, - 'flush_immediately': {'key': 'flushImmediately', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDataExplorerSink, self).__init__(**kwargs) - self.type = 'AzureDataExplorerSink' # type: str - self.ingestion_mapping_name = kwargs.get('ingestion_mapping_name', None) - self.ingestion_mapping_as_json = kwargs.get('ingestion_mapping_as_json', None) - self.flush_immediately = kwargs.get('flush_immediately', None) - - -class AzureDataExplorerSource(CopySource): - """A copy activity Azure Data Explorer (Kusto) source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: Required. Database query. Should be a Kusto Query Language (KQL) query. Type: - string (or Expression with resultType string). - :type query: any - :param no_truncation: The name of the Boolean option that controls whether truncation is - applied to result-sets that go beyond a certain row-count limit. - :type no_truncation: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])).. - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - 'query': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'no_truncation': {'key': 'noTruncation', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDataExplorerSource, self).__init__(**kwargs) - self.type = 'AzureDataExplorerSource' # type: str - self.query = kwargs['query'] - self.no_truncation = kwargs.get('no_truncation', None) - self.query_timeout = kwargs.get('query_timeout', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class AzureDataExplorerTableDataset(Dataset): - """The Azure Data Explorer (Kusto) dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table: The table name of the Azure Data Explorer database. Type: string (or Expression - with resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDataExplorerTableDataset, self).__init__(**kwargs) - self.type = 'AzureDataExplorerTable' # type: str - self.table = kwargs.get('table', None) - - -class AzureDataLakeAnalyticsLinkedService(LinkedService): - """Azure Data Lake Analytics linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param account_name: Required. The Azure Data Lake Analytics account name. Type: string (or - Expression with resultType string). - :type account_name: any - :param service_principal_id: The ID of the application used to authenticate against the Azure - Data Lake Analytics account. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: The Key of the application used to authenticate against the Azure - Data Lake Analytics account. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: Required. The name or ID of the tenant to which the service principal belongs. - Type: string (or Expression with resultType string). - :type tenant: any - :param subscription_id: Data Lake Analytics account subscription ID (if different from Data - Factory account). Type: string (or Expression with resultType string). - :type subscription_id: any - :param resource_group_name: Data Lake Analytics account resource group name (if different from - Data Factory account). Type: string (or Expression with resultType string). - :type resource_group_name: any - :param data_lake_analytics_uri: Azure Data Lake Analytics URI Type: string (or Expression with - resultType string). - :type data_lake_analytics_uri: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'account_name': {'required': True}, - 'tenant': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'account_name': {'key': 'typeProperties.accountName', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'subscription_id': {'key': 'typeProperties.subscriptionId', 'type': 'object'}, - 'resource_group_name': {'key': 'typeProperties.resourceGroupName', 'type': 'object'}, - 'data_lake_analytics_uri': {'key': 'typeProperties.dataLakeAnalyticsUri', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDataLakeAnalyticsLinkedService, self).__init__(**kwargs) - self.type = 'AzureDataLakeAnalytics' # type: str - self.account_name = kwargs['account_name'] - self.service_principal_id = kwargs.get('service_principal_id', None) - self.service_principal_key = kwargs.get('service_principal_key', None) - self.tenant = kwargs['tenant'] - self.subscription_id = kwargs.get('subscription_id', None) - self.resource_group_name = kwargs.get('resource_group_name', None) - self.data_lake_analytics_uri = kwargs.get('data_lake_analytics_uri', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AzureDataLakeStoreDataset(Dataset): - """Azure Data Lake Store dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param folder_path: Path to the folder in the Azure Data Lake Store. Type: string (or - Expression with resultType string). - :type folder_path: any - :param file_name: The name of the file in the Azure Data Lake Store. Type: string (or - Expression with resultType string). - :type file_name: any - :param format: The format of the Data Lake Store. - :type format: ~azure.synapse.artifacts.v2020_12_01.models.DatasetStorageFormat - :param compression: The data compression method used for the item(s) in the Azure Data Lake - Store. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, - 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, - 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDataLakeStoreDataset, self).__init__(**kwargs) - self.type = 'AzureDataLakeStoreFile' # type: str - self.folder_path = kwargs.get('folder_path', None) - self.file_name = kwargs.get('file_name', None) - self.format = kwargs.get('format', None) - self.compression = kwargs.get('compression', None) - - -class AzureDataLakeStoreLinkedService(LinkedService): - """Azure Data Lake Store linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param data_lake_store_uri: Required. Data Lake Store service URI. Type: string (or Expression - with resultType string). - :type data_lake_store_uri: any - :param service_principal_id: The ID of the application used to authenticate against the Azure - Data Lake Store account. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: The Key of the application used to authenticate against the Azure - Data Lake Store account. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: The name or ID of the tenant to which the service principal belongs. Type: - string (or Expression with resultType string). - :type tenant: any - :param azure_cloud_type: Indicates the azure cloud type of the service principle auth. Allowed - values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data - factory regions’ cloud type. Type: string (or Expression with resultType string). - :type azure_cloud_type: any - :param account_name: Data Lake Store account name. Type: string (or Expression with resultType - string). - :type account_name: any - :param subscription_id: Data Lake Store account subscription ID (if different from Data Factory - account). Type: string (or Expression with resultType string). - :type subscription_id: any - :param resource_group_name: Data Lake Store account resource group name (if different from Data - Factory account). Type: string (or Expression with resultType string). - :type resource_group_name: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'data_lake_store_uri': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'data_lake_store_uri': {'key': 'typeProperties.dataLakeStoreUri', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'azure_cloud_type': {'key': 'typeProperties.azureCloudType', 'type': 'object'}, - 'account_name': {'key': 'typeProperties.accountName', 'type': 'object'}, - 'subscription_id': {'key': 'typeProperties.subscriptionId', 'type': 'object'}, - 'resource_group_name': {'key': 'typeProperties.resourceGroupName', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDataLakeStoreLinkedService, self).__init__(**kwargs) - self.type = 'AzureDataLakeStore' # type: str - self.data_lake_store_uri = kwargs['data_lake_store_uri'] - self.service_principal_id = kwargs.get('service_principal_id', None) - self.service_principal_key = kwargs.get('service_principal_key', None) - self.tenant = kwargs.get('tenant', None) - self.azure_cloud_type = kwargs.get('azure_cloud_type', None) - self.account_name = kwargs.get('account_name', None) - self.subscription_id = kwargs.get('subscription_id', None) - self.resource_group_name = kwargs.get('resource_group_name', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AzureDataLakeStoreLocation(DatasetLocation): - """The location of azure data lake store dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDataLakeStoreLocation, self).__init__(**kwargs) - self.type = 'AzureDataLakeStoreLocation' # type: str - - -class AzureDataLakeStoreReadSettings(StoreReadSettings): - """Azure data lake store read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: ADLS wildcardFolderPath. Type: string (or Expression with - resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: ADLS wildcardFileName. Type: string (or Expression with resultType - string). - :type wildcard_file_name: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param list_after: Lists files after the value (exclusive) based on file/folder names’ - lexicographical order. Applies under the folderPath in data set, and filter files/sub-folders - under the folderPath. Type: string (or Expression with resultType string). - :type list_after: any - :param list_before: Lists files before the value (inclusive) based on file/folder names’ - lexicographical order. Applies under the folderPath in data set, and filter files/sub-folders - under the folderPath. Type: string (or Expression with resultType string). - :type list_before: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'list_after': {'key': 'listAfter', 'type': 'object'}, - 'list_before': {'key': 'listBefore', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - 'modified_datetime_start': {'key': 'modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'modifiedDatetimeEnd', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDataLakeStoreReadSettings, self).__init__(**kwargs) - self.type = 'AzureDataLakeStoreReadSettings' # type: str - self.recursive = kwargs.get('recursive', None) - self.wildcard_folder_path = kwargs.get('wildcard_folder_path', None) - self.wildcard_file_name = kwargs.get('wildcard_file_name', None) - self.file_list_path = kwargs.get('file_list_path', None) - self.list_after = kwargs.get('list_after', None) - self.list_before = kwargs.get('list_before', None) - self.enable_partition_discovery = kwargs.get('enable_partition_discovery', None) - self.partition_root_path = kwargs.get('partition_root_path', None) - self.delete_files_after_completion = kwargs.get('delete_files_after_completion', None) - self.modified_datetime_start = kwargs.get('modified_datetime_start', None) - self.modified_datetime_end = kwargs.get('modified_datetime_end', None) - - -class AzureDataLakeStoreSink(CopySink): - """A copy activity Azure Data Lake Store sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - :param enable_adls_single_file_parallel: Single File Parallel. - :type enable_adls_single_file_parallel: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - 'enable_adls_single_file_parallel': {'key': 'enableAdlsSingleFileParallel', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDataLakeStoreSink, self).__init__(**kwargs) - self.type = 'AzureDataLakeStoreSink' # type: str - self.copy_behavior = kwargs.get('copy_behavior', None) - self.enable_adls_single_file_parallel = kwargs.get('enable_adls_single_file_parallel', None) - - -class AzureDataLakeStoreSource(CopySource): - """A copy activity Azure Data Lake source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDataLakeStoreSource, self).__init__(**kwargs) - self.type = 'AzureDataLakeStoreSource' # type: str - self.recursive = kwargs.get('recursive', None) - - -class AzureDataLakeStoreWriteSettings(StoreWriteSettings): - """Azure data lake store write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - :param expiry_date_time: Specifies the expiry time of the written files. The time is applied to - the UTC time zone in the format of "2018-12-01T05:00:00Z". Default value is NULL. Type: integer - (or Expression with resultType integer). - :type expiry_date_time: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - 'expiry_date_time': {'key': 'expiryDateTime', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureDataLakeStoreWriteSettings, self).__init__(**kwargs) - self.type = 'AzureDataLakeStoreWriteSettings' # type: str - self.expiry_date_time = kwargs.get('expiry_date_time', None) - - -class Resource(msrest.serialization.Model): - """Common fields that are returned in the response for all Azure Resource Manager resources. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(Resource, self).__init__(**kwargs) - self.id = None - self.name = None - self.type = None - - -class AzureEntityResource(Resource): - """The resource model definition for an Azure Resource Manager resource with an etag. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureEntityResource, self).__init__(**kwargs) - self.etag = None - - -class AzureFileStorageLinkedService(LinkedService): - """Azure File Storage linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. Host name of the server. Type: string (or Expression with resultType - string). - :type host: any - :param user_id: User ID to logon the server. Type: string (or Expression with resultType - string). - :type user_id: any - :param password: Password to logon the server. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param connection_string: The connection string. It is mutually exclusive with sasUri property. - Type: string, SecureString or AzureKeyVaultSecretReference. - :type connection_string: any - :param account_key: The Azure key vault secret reference of accountKey in connection string. - :type account_key: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param sas_uri: SAS URI of the Azure File resource. It is mutually exclusive with - connectionString property. Type: string, SecureString or AzureKeyVaultSecretReference. - :type sas_uri: any - :param sas_token: The Azure key vault secret reference of sasToken in sas uri. - :type sas_token: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param file_share: The azure file share name. It is required when auth with - accountKey/sasToken. Type: string (or Expression with resultType string). - :type file_share: any - :param snapshot: The azure file share snapshot version. Type: string (or Expression with - resultType string). - :type snapshot: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'user_id': {'key': 'typeProperties.userId', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'account_key': {'key': 'typeProperties.accountKey', 'type': 'AzureKeyVaultSecretReference'}, - 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'object'}, - 'sas_token': {'key': 'typeProperties.sasToken', 'type': 'AzureKeyVaultSecretReference'}, - 'file_share': {'key': 'typeProperties.fileShare', 'type': 'object'}, - 'snapshot': {'key': 'typeProperties.snapshot', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureFileStorageLinkedService, self).__init__(**kwargs) - self.type = 'AzureFileStorage' # type: str - self.host = kwargs['host'] - self.user_id = kwargs.get('user_id', None) - self.password = kwargs.get('password', None) - self.connection_string = kwargs.get('connection_string', None) - self.account_key = kwargs.get('account_key', None) - self.sas_uri = kwargs.get('sas_uri', None) - self.sas_token = kwargs.get('sas_token', None) - self.file_share = kwargs.get('file_share', None) - self.snapshot = kwargs.get('snapshot', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AzureFileStorageLocation(DatasetLocation): - """The location of file server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureFileStorageLocation, self).__init__(**kwargs) - self.type = 'AzureFileStorageLocation' # type: str - - -class AzureFileStorageReadSettings(StoreReadSettings): - """Azure File Storage read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: Azure File Storage wildcardFolderPath. Type: string (or Expression - with resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: Azure File Storage wildcardFileName. Type: string (or Expression - with resultType string). - :type wildcard_file_name: any - :param prefix: The prefix filter for the Azure File name starting from root path. Type: string - (or Expression with resultType string). - :type prefix: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'prefix': {'key': 'prefix', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - 'modified_datetime_start': {'key': 'modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'modifiedDatetimeEnd', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureFileStorageReadSettings, self).__init__(**kwargs) - self.type = 'AzureFileStorageReadSettings' # type: str - self.recursive = kwargs.get('recursive', None) - self.wildcard_folder_path = kwargs.get('wildcard_folder_path', None) - self.wildcard_file_name = kwargs.get('wildcard_file_name', None) - self.prefix = kwargs.get('prefix', None) - self.file_list_path = kwargs.get('file_list_path', None) - self.enable_partition_discovery = kwargs.get('enable_partition_discovery', None) - self.partition_root_path = kwargs.get('partition_root_path', None) - self.delete_files_after_completion = kwargs.get('delete_files_after_completion', None) - self.modified_datetime_start = kwargs.get('modified_datetime_start', None) - self.modified_datetime_end = kwargs.get('modified_datetime_end', None) - - -class AzureFileStorageWriteSettings(StoreWriteSettings): - """Azure File Storage write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureFileStorageWriteSettings, self).__init__(**kwargs) - self.type = 'AzureFileStorageWriteSettings' # type: str - - -class AzureFunctionActivity(ExecutionActivity): - """Azure Function activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param method: Required. Rest API method for target endpoint. Possible values include: "GET", - "POST", "PUT", "DELETE", "OPTIONS", "HEAD", "TRACE". - :type method: str or ~azure.synapse.artifacts.v2020_12_01.models.AzureFunctionActivityMethod - :param function_name: Required. Name of the Function that the Azure Function Activity will - call. Type: string (or Expression with resultType string). - :type function_name: any - :param headers: Represents the headers that will be sent to the request. For example, to set - the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": - "application/json" }. Type: string (or Expression with resultType string). - :type headers: any - :param body: Represents the payload that will be sent to the endpoint. Required for POST/PUT - method, not allowed for GET method Type: string (or Expression with resultType string). - :type body: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'method': {'required': True}, - 'function_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'method': {'key': 'typeProperties.method', 'type': 'str'}, - 'function_name': {'key': 'typeProperties.functionName', 'type': 'object'}, - 'headers': {'key': 'typeProperties.headers', 'type': 'object'}, - 'body': {'key': 'typeProperties.body', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureFunctionActivity, self).__init__(**kwargs) - self.type = 'AzureFunctionActivity' # type: str - self.method = kwargs['method'] - self.function_name = kwargs['function_name'] - self.headers = kwargs.get('headers', None) - self.body = kwargs.get('body', None) - - -class AzureFunctionLinkedService(LinkedService): - """Azure Function linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param function_app_url: Required. The endpoint of the Azure Function App. URL will be in the - format https://:code:``.azurewebsites.net. - :type function_app_url: any - :param function_key: Function or Host key for Azure Function App. - :type function_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'function_app_url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'function_app_url': {'key': 'typeProperties.functionAppUrl', 'type': 'object'}, - 'function_key': {'key': 'typeProperties.functionKey', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureFunctionLinkedService, self).__init__(**kwargs) - self.type = 'AzureFunction' # type: str - self.function_app_url = kwargs['function_app_url'] - self.function_key = kwargs.get('function_key', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AzureKeyVaultLinkedService(LinkedService): - """Azure Key Vault linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param base_url: Required. The base URL of the Azure Key Vault. e.g. - https://myakv.vault.azure.net Type: string (or Expression with resultType string). - :type base_url: any - """ - - _validation = { - 'type': {'required': True}, - 'base_url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'base_url': {'key': 'typeProperties.baseUrl', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureKeyVaultLinkedService, self).__init__(**kwargs) - self.type = 'AzureKeyVault' # type: str - self.base_url = kwargs['base_url'] - - -class SecretBase(msrest.serialization.Model): - """The base definition of a secret type. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AzureKeyVaultSecretReference, SecureString. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Type of the secret.Constant filled by server. - :type type: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - } - - _subtype_map = { - 'type': {'AzureKeyVaultSecret': 'AzureKeyVaultSecretReference', 'SecureString': 'SecureString'} - } - - def __init__( - self, - **kwargs - ): - super(SecretBase, self).__init__(**kwargs) - self.type = None # type: Optional[str] - - -class AzureKeyVaultSecretReference(SecretBase): - """Azure Key Vault secret reference. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Type of the secret.Constant filled by server. - :type type: str - :param store: Required. The Azure Key Vault linked service reference. - :type store: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param secret_name: Required. The name of the secret in Azure Key Vault. Type: string (or - Expression with resultType string). - :type secret_name: any - :param secret_version: The version of the secret in Azure Key Vault. The default value is the - latest version of the secret. Type: string (or Expression with resultType string). - :type secret_version: any - """ - - _validation = { - 'type': {'required': True}, - 'store': {'required': True}, - 'secret_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'store': {'key': 'store', 'type': 'LinkedServiceReference'}, - 'secret_name': {'key': 'secretName', 'type': 'object'}, - 'secret_version': {'key': 'secretVersion', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureKeyVaultSecretReference, self).__init__(**kwargs) - self.type = 'AzureKeyVaultSecret' # type: str - self.store = kwargs['store'] - self.secret_name = kwargs['secret_name'] - self.secret_version = kwargs.get('secret_version', None) - - -class AzureMariaDBLinkedService(LinkedService): - """Azure Database for MariaDB linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: An ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureMariaDBLinkedService, self).__init__(**kwargs) - self.type = 'AzureMariaDB' # type: str - self.connection_string = kwargs.get('connection_string', None) - self.pwd = kwargs.get('pwd', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AzureMariaDBSource(TabularSource): - """A copy activity Azure MariaDB source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureMariaDBSource, self).__init__(**kwargs) - self.type = 'AzureMariaDBSource' # type: str - self.query = kwargs.get('query', None) - - -class AzureMariaDBTableDataset(Dataset): - """Azure Database for MariaDB dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureMariaDBTableDataset, self).__init__(**kwargs) - self.type = 'AzureMariaDBTable' # type: str - self.table_name = kwargs.get('table_name', None) - - -class AzureMLBatchExecutionActivity(ExecutionActivity): - """Azure ML Batch Execution activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param global_parameters: Key,Value pairs to be passed to the Azure ML Batch Execution Service - endpoint. Keys must match the names of web service parameters defined in the published Azure ML - web service. Values will be passed in the GlobalParameters property of the Azure ML batch - execution request. - :type global_parameters: dict[str, any] - :param web_service_outputs: Key,Value pairs, mapping the names of Azure ML endpoint's Web - Service Outputs to AzureMLWebServiceFile objects specifying the output Blob locations. This - information will be passed in the WebServiceOutputs property of the Azure ML batch execution - request. - :type web_service_outputs: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.AzureMLWebServiceFile] - :param web_service_inputs: Key,Value pairs, mapping the names of Azure ML endpoint's Web - Service Inputs to AzureMLWebServiceFile objects specifying the input Blob locations.. This - information will be passed in the WebServiceInputs property of the Azure ML batch execution - request. - :type web_service_inputs: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.AzureMLWebServiceFile] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'global_parameters': {'key': 'typeProperties.globalParameters', 'type': '{object}'}, - 'web_service_outputs': {'key': 'typeProperties.webServiceOutputs', 'type': '{AzureMLWebServiceFile}'}, - 'web_service_inputs': {'key': 'typeProperties.webServiceInputs', 'type': '{AzureMLWebServiceFile}'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureMLBatchExecutionActivity, self).__init__(**kwargs) - self.type = 'AzureMLBatchExecution' # type: str - self.global_parameters = kwargs.get('global_parameters', None) - self.web_service_outputs = kwargs.get('web_service_outputs', None) - self.web_service_inputs = kwargs.get('web_service_inputs', None) - - -class AzureMLExecutePipelineActivity(ExecutionActivity): - """Azure ML Execute Pipeline activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param ml_pipeline_id: Required. ID of the published Azure ML pipeline. Type: string (or - Expression with resultType string). - :type ml_pipeline_id: any - :param experiment_name: Run history experiment name of the pipeline run. This information will - be passed in the ExperimentName property of the published pipeline execution request. Type: - string (or Expression with resultType string). - :type experiment_name: any - :param ml_pipeline_parameters: Key,Value pairs to be passed to the published Azure ML pipeline - endpoint. Keys must match the names of pipeline parameters defined in the published pipeline. - Values will be passed in the ParameterAssignments property of the published pipeline execution - request. Type: object with key value pairs (or Expression with resultType object). - :type ml_pipeline_parameters: any - :param ml_parent_run_id: The parent Azure ML Service pipeline run id. This information will be - passed in the ParentRunId property of the published pipeline execution request. Type: string - (or Expression with resultType string). - :type ml_parent_run_id: any - :param continue_on_step_failure: Whether to continue execution of other steps in the - PipelineRun if a step fails. This information will be passed in the continueOnStepFailure - property of the published pipeline execution request. Type: boolean (or Expression with - resultType boolean). - :type continue_on_step_failure: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'ml_pipeline_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'ml_pipeline_id': {'key': 'typeProperties.mlPipelineId', 'type': 'object'}, - 'experiment_name': {'key': 'typeProperties.experimentName', 'type': 'object'}, - 'ml_pipeline_parameters': {'key': 'typeProperties.mlPipelineParameters', 'type': 'object'}, - 'ml_parent_run_id': {'key': 'typeProperties.mlParentRunId', 'type': 'object'}, - 'continue_on_step_failure': {'key': 'typeProperties.continueOnStepFailure', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureMLExecutePipelineActivity, self).__init__(**kwargs) - self.type = 'AzureMLExecutePipeline' # type: str - self.ml_pipeline_id = kwargs['ml_pipeline_id'] - self.experiment_name = kwargs.get('experiment_name', None) - self.ml_pipeline_parameters = kwargs.get('ml_pipeline_parameters', None) - self.ml_parent_run_id = kwargs.get('ml_parent_run_id', None) - self.continue_on_step_failure = kwargs.get('continue_on_step_failure', None) - - -class AzureMLLinkedService(LinkedService): - """Azure ML Studio Web Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param ml_endpoint: Required. The Batch Execution REST URL for an Azure ML Studio Web Service - endpoint. Type: string (or Expression with resultType string). - :type ml_endpoint: any - :param api_key: Required. The API key for accessing the Azure ML model endpoint. - :type api_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param update_resource_endpoint: The Update Resource REST URL for an Azure ML Studio Web - Service endpoint. Type: string (or Expression with resultType string). - :type update_resource_endpoint: any - :param service_principal_id: The ID of the service principal used to authenticate against the - ARM-based updateResourceEndpoint of an Azure ML Studio web service. Type: string (or Expression - with resultType string). - :type service_principal_id: any - :param service_principal_key: The key of the service principal used to authenticate against the - ARM-based updateResourceEndpoint of an Azure ML Studio web service. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: The name or ID of the tenant to which the service principal belongs. Type: - string (or Expression with resultType string). - :type tenant: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'ml_endpoint': {'required': True}, - 'api_key': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'ml_endpoint': {'key': 'typeProperties.mlEndpoint', 'type': 'object'}, - 'api_key': {'key': 'typeProperties.apiKey', 'type': 'SecretBase'}, - 'update_resource_endpoint': {'key': 'typeProperties.updateResourceEndpoint', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureMLLinkedService, self).__init__(**kwargs) - self.type = 'AzureML' # type: str - self.ml_endpoint = kwargs['ml_endpoint'] - self.api_key = kwargs['api_key'] - self.update_resource_endpoint = kwargs.get('update_resource_endpoint', None) - self.service_principal_id = kwargs.get('service_principal_id', None) - self.service_principal_key = kwargs.get('service_principal_key', None) - self.tenant = kwargs.get('tenant', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AzureMLServiceLinkedService(LinkedService): - """Azure ML Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param subscription_id: Required. Azure ML Service workspace subscription ID. Type: string (or - Expression with resultType string). - :type subscription_id: any - :param resource_group_name: Required. Azure ML Service workspace resource group name. Type: - string (or Expression with resultType string). - :type resource_group_name: any - :param ml_workspace_name: Required. Azure ML Service workspace name. Type: string (or - Expression with resultType string). - :type ml_workspace_name: any - :param service_principal_id: The ID of the service principal used to authenticate against the - endpoint of a published Azure ML Service pipeline. Type: string (or Expression with resultType - string). - :type service_principal_id: any - :param service_principal_key: The key of the service principal used to authenticate against the - endpoint of a published Azure ML Service pipeline. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: The name or ID of the tenant to which the service principal belongs. Type: - string (or Expression with resultType string). - :type tenant: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'subscription_id': {'required': True}, - 'resource_group_name': {'required': True}, - 'ml_workspace_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'subscription_id': {'key': 'typeProperties.subscriptionId', 'type': 'object'}, - 'resource_group_name': {'key': 'typeProperties.resourceGroupName', 'type': 'object'}, - 'ml_workspace_name': {'key': 'typeProperties.mlWorkspaceName', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureMLServiceLinkedService, self).__init__(**kwargs) - self.type = 'AzureMLService' # type: str - self.subscription_id = kwargs['subscription_id'] - self.resource_group_name = kwargs['resource_group_name'] - self.ml_workspace_name = kwargs['ml_workspace_name'] - self.service_principal_id = kwargs.get('service_principal_id', None) - self.service_principal_key = kwargs.get('service_principal_key', None) - self.tenant = kwargs.get('tenant', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AzureMLUpdateResourceActivity(ExecutionActivity): - """Azure ML Update Resource management activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param trained_model_name: Required. Name of the Trained Model module in the Web Service - experiment to be updated. Type: string (or Expression with resultType string). - :type trained_model_name: any - :param trained_model_linked_service_name: Required. Name of Azure Storage linked service - holding the .ilearner file that will be uploaded by the update operation. - :type trained_model_linked_service_name: - ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param trained_model_file_path: Required. The relative file path in trainedModelLinkedService - to represent the .ilearner file that will be uploaded by the update operation. Type: string - (or Expression with resultType string). - :type trained_model_file_path: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'trained_model_name': {'required': True}, - 'trained_model_linked_service_name': {'required': True}, - 'trained_model_file_path': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'trained_model_name': {'key': 'typeProperties.trainedModelName', 'type': 'object'}, - 'trained_model_linked_service_name': {'key': 'typeProperties.trainedModelLinkedServiceName', 'type': 'LinkedServiceReference'}, - 'trained_model_file_path': {'key': 'typeProperties.trainedModelFilePath', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureMLUpdateResourceActivity, self).__init__(**kwargs) - self.type = 'AzureMLUpdateResource' # type: str - self.trained_model_name = kwargs['trained_model_name'] - self.trained_model_linked_service_name = kwargs['trained_model_linked_service_name'] - self.trained_model_file_path = kwargs['trained_model_file_path'] - - -class AzureMLWebServiceFile(msrest.serialization.Model): - """Azure ML WebService Input/Output file. - - All required parameters must be populated in order to send to Azure. - - :param file_path: Required. The relative file path, including container name, in the Azure Blob - Storage specified by the LinkedService. Type: string (or Expression with resultType string). - :type file_path: any - :param linked_service_name: Required. Reference to an Azure Storage LinkedService, where Azure - ML WebService Input/Output file located. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - """ - - _validation = { - 'file_path': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'file_path': {'key': 'filePath', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureMLWebServiceFile, self).__init__(**kwargs) - self.file_path = kwargs['file_path'] - self.linked_service_name = kwargs['linked_service_name'] - - -class AzureMySqlLinkedService(LinkedService): - """Azure MySQL database linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param password: The Azure key vault secret reference of password in connection string. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureMySqlLinkedService, self).__init__(**kwargs) - self.type = 'AzureMySql' # type: str - self.connection_string = kwargs['connection_string'] - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AzureMySqlSink(CopySink): - """A copy activity Azure MySql sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param pre_copy_script: A query to execute before starting the copy. Type: string (or - Expression with resultType string). - :type pre_copy_script: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureMySqlSink, self).__init__(**kwargs) - self.type = 'AzureMySqlSink' # type: str - self.pre_copy_script = kwargs.get('pre_copy_script', None) - - -class AzureMySqlSource(TabularSource): - """A copy activity Azure MySQL source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureMySqlSource, self).__init__(**kwargs) - self.type = 'AzureMySqlSource' # type: str - self.query = kwargs.get('query', None) - - -class AzureMySqlTableDataset(Dataset): - """The Azure MySQL database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The Azure MySQL database table name. Type: string (or Expression with - resultType string). - :type table_name: any - :param table: The name of Azure MySQL database table. Type: string (or Expression with - resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureMySqlTableDataset, self).__init__(**kwargs) - self.type = 'AzureMySqlTable' # type: str - self.table_name = kwargs.get('table_name', None) - self.table = kwargs.get('table', None) - - -class AzurePostgreSqlLinkedService(LinkedService): - """Azure PostgreSQL linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: An ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param password: The Azure key vault secret reference of password in connection string. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzurePostgreSqlLinkedService, self).__init__(**kwargs) - self.type = 'AzurePostgreSql' # type: str - self.connection_string = kwargs.get('connection_string', None) - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AzurePostgreSqlSink(CopySink): - """A copy activity Azure PostgreSQL sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param pre_copy_script: A query to execute before starting the copy. Type: string (or - Expression with resultType string). - :type pre_copy_script: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzurePostgreSqlSink, self).__init__(**kwargs) - self.type = 'AzurePostgreSqlSink' # type: str - self.pre_copy_script = kwargs.get('pre_copy_script', None) - - -class AzurePostgreSqlSource(TabularSource): - """A copy activity Azure PostgreSQL source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzurePostgreSqlSource, self).__init__(**kwargs) - self.type = 'AzurePostgreSqlSource' # type: str - self.query = kwargs.get('query', None) - - -class AzurePostgreSqlTableDataset(Dataset): - """Azure PostgreSQL dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name of the Azure PostgreSQL database which includes both schema - and table. Type: string (or Expression with resultType string). - :type table_name: any - :param table: The table name of the Azure PostgreSQL database. Type: string (or Expression with - resultType string). - :type table: any - :param schema_type_properties_schema: The schema name of the Azure PostgreSQL database. Type: - string (or Expression with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzurePostgreSqlTableDataset, self).__init__(**kwargs) - self.type = 'AzurePostgreSqlTable' # type: str - self.table_name = kwargs.get('table_name', None) - self.table = kwargs.get('table', None) - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - - -class AzureQueueSink(CopySink): - """A copy activity Azure Queue sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureQueueSink, self).__init__(**kwargs) - self.type = 'AzureQueueSink' # type: str - - -class AzureSearchIndexDataset(Dataset): - """The Azure Search Index. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param index_name: Required. The name of the Azure Search Index. Type: string (or Expression - with resultType string). - :type index_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'index_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'index_name': {'key': 'typeProperties.indexName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureSearchIndexDataset, self).__init__(**kwargs) - self.type = 'AzureSearchIndex' # type: str - self.index_name = kwargs['index_name'] - - -class AzureSearchIndexSink(CopySink): - """A copy activity Azure Search Index sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param write_behavior: Specify the write behavior when upserting documents into Azure Search - Index. Possible values include: "Merge", "Upload". - :type write_behavior: str or - ~azure.synapse.artifacts.v2020_12_01.models.AzureSearchIndexWriteBehaviorType - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureSearchIndexSink, self).__init__(**kwargs) - self.type = 'AzureSearchIndexSink' # type: str - self.write_behavior = kwargs.get('write_behavior', None) - - -class AzureSearchLinkedService(LinkedService): - """Linked service for Windows Azure Search Service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param url: Required. URL for Azure Search service. Type: string (or Expression with resultType - string). - :type url: any - :param key: Admin Key for Azure Search service. - :type key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'key': {'key': 'typeProperties.key', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureSearchLinkedService, self).__init__(**kwargs) - self.type = 'AzureSearch' # type: str - self.url = kwargs['url'] - self.key = kwargs.get('key', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AzureSqlDatabaseLinkedService(LinkedService): - """Microsoft Azure SQL Database linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param password: The Azure key vault secret reference of password in connection string. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param service_principal_id: The ID of the service principal used to authenticate against Azure - SQL Database. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: The key of the service principal used to authenticate against - Azure SQL Database. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: The name or ID of the tenant to which the service principal belongs. Type: - string (or Expression with resultType string). - :type tenant: any - :param azure_cloud_type: Indicates the azure cloud type of the service principle auth. Allowed - values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data - factory regions’ cloud type. Type: string (or Expression with resultType string). - :type azure_cloud_type: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'azure_cloud_type': {'key': 'typeProperties.azureCloudType', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureSqlDatabaseLinkedService, self).__init__(**kwargs) - self.type = 'AzureSqlDatabase' # type: str - self.connection_string = kwargs['connection_string'] - self.password = kwargs.get('password', None) - self.service_principal_id = kwargs.get('service_principal_id', None) - self.service_principal_key = kwargs.get('service_principal_key', None) - self.tenant = kwargs.get('tenant', None) - self.azure_cloud_type = kwargs.get('azure_cloud_type', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AzureSqlDWLinkedService(LinkedService): - """Azure SQL Data Warehouse linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. Type: string, SecureString or AzureKeyVaultSecretReference. - :type connection_string: any - :param password: The Azure key vault secret reference of password in connection string. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param service_principal_id: The ID of the service principal used to authenticate against Azure - SQL Data Warehouse. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: The key of the service principal used to authenticate against - Azure SQL Data Warehouse. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: The name or ID of the tenant to which the service principal belongs. Type: - string (or Expression with resultType string). - :type tenant: any - :param azure_cloud_type: Indicates the azure cloud type of the service principle auth. Allowed - values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data - factory regions’ cloud type. Type: string (or Expression with resultType string). - :type azure_cloud_type: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'azure_cloud_type': {'key': 'typeProperties.azureCloudType', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureSqlDWLinkedService, self).__init__(**kwargs) - self.type = 'AzureSqlDW' # type: str - self.connection_string = kwargs['connection_string'] - self.password = kwargs.get('password', None) - self.service_principal_id = kwargs.get('service_principal_id', None) - self.service_principal_key = kwargs.get('service_principal_key', None) - self.tenant = kwargs.get('tenant', None) - self.azure_cloud_type = kwargs.get('azure_cloud_type', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AzureSqlDWTableDataset(Dataset): - """The Azure SQL Data Warehouse dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param schema_type_properties_schema: The schema name of the Azure SQL Data Warehouse. Type: - string (or Expression with resultType string). - :type schema_type_properties_schema: any - :param table: The table name of the Azure SQL Data Warehouse. Type: string (or Expression with - resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureSqlDWTableDataset, self).__init__(**kwargs) - self.type = 'AzureSqlDWTable' # type: str - self.table_name = kwargs.get('table_name', None) - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - self.table = kwargs.get('table', None) - - -class AzureSqlMILinkedService(LinkedService): - """Azure SQL Managed Instance linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param password: The Azure key vault secret reference of password in connection string. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param service_principal_id: The ID of the service principal used to authenticate against Azure - SQL Managed Instance. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: The key of the service principal used to authenticate against - Azure SQL Managed Instance. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: The name or ID of the tenant to which the service principal belongs. Type: - string (or Expression with resultType string). - :type tenant: any - :param azure_cloud_type: Indicates the azure cloud type of the service principle auth. Allowed - values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data - factory regions’ cloud type. Type: string (or Expression with resultType string). - :type azure_cloud_type: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'azure_cloud_type': {'key': 'typeProperties.azureCloudType', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureSqlMILinkedService, self).__init__(**kwargs) - self.type = 'AzureSqlMI' # type: str - self.connection_string = kwargs['connection_string'] - self.password = kwargs.get('password', None) - self.service_principal_id = kwargs.get('service_principal_id', None) - self.service_principal_key = kwargs.get('service_principal_key', None) - self.tenant = kwargs.get('tenant', None) - self.azure_cloud_type = kwargs.get('azure_cloud_type', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AzureSqlMITableDataset(Dataset): - """The Azure SQL Managed Instance dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param schema_type_properties_schema: The schema name of the Azure SQL Managed Instance. Type: - string (or Expression with resultType string). - :type schema_type_properties_schema: any - :param table: The table name of the Azure SQL Managed Instance dataset. Type: string (or - Expression with resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureSqlMITableDataset, self).__init__(**kwargs) - self.type = 'AzureSqlMITable' # type: str - self.table_name = kwargs.get('table_name', None) - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - self.table = kwargs.get('table', None) - - -class AzureSqlSink(CopySink): - """A copy activity Azure SQL sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param sql_writer_stored_procedure_name: SQL writer stored procedure name. Type: string (or - Expression with resultType string). - :type sql_writer_stored_procedure_name: any - :param sql_writer_table_type: SQL writer table type. Type: string (or Expression with - resultType string). - :type sql_writer_table_type: any - :param pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType - string). - :type pre_copy_script: any - :param stored_procedure_parameters: SQL stored procedure parameters. - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - :param stored_procedure_table_type_parameter_name: The stored procedure parameter name of the - table type. Type: string (or Expression with resultType string). - :type stored_procedure_table_type_parameter_name: any - :param table_option: The option to handle sink table, such as autoCreate. For now only - 'autoCreate' value is supported. Type: string (or Expression with resultType string). - :type table_option: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'sql_writer_stored_procedure_name': {'key': 'sqlWriterStoredProcedureName', 'type': 'object'}, - 'sql_writer_table_type': {'key': 'sqlWriterTableType', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - 'stored_procedure_table_type_parameter_name': {'key': 'storedProcedureTableTypeParameterName', 'type': 'object'}, - 'table_option': {'key': 'tableOption', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureSqlSink, self).__init__(**kwargs) - self.type = 'AzureSqlSink' # type: str - self.sql_writer_stored_procedure_name = kwargs.get('sql_writer_stored_procedure_name', None) - self.sql_writer_table_type = kwargs.get('sql_writer_table_type', None) - self.pre_copy_script = kwargs.get('pre_copy_script', None) - self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) - self.stored_procedure_table_type_parameter_name = kwargs.get('stored_procedure_table_type_parameter_name', None) - self.table_option = kwargs.get('table_option', None) - - -class AzureSqlSource(TabularSource): - """A copy activity Azure SQL source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). - :type sql_reader_query: any - :param sql_reader_stored_procedure_name: Name of the stored procedure for a SQL Database - source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression - with resultType string). - :type sql_reader_stored_procedure_name: any - :param stored_procedure_parameters: Value and type setting for stored procedure parameters. - Example: "{Parameter1: {value: "1", type: "int"}}". - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - :param produce_additional_types: Which additional types to produce. - :type produce_additional_types: any - :param partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". - :type partition_option: any - :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.SqlPartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, - 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - 'produce_additional_types': {'key': 'produceAdditionalTypes', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'object'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureSqlSource, self).__init__(**kwargs) - self.type = 'AzureSqlSource' # type: str - self.sql_reader_query = kwargs.get('sql_reader_query', None) - self.sql_reader_stored_procedure_name = kwargs.get('sql_reader_stored_procedure_name', None) - self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) - self.produce_additional_types = kwargs.get('produce_additional_types', None) - self.partition_option = kwargs.get('partition_option', None) - self.partition_settings = kwargs.get('partition_settings', None) - - -class AzureSqlTableDataset(Dataset): - """The Azure SQL Server database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param schema_type_properties_schema: The schema name of the Azure SQL database. Type: string - (or Expression with resultType string). - :type schema_type_properties_schema: any - :param table: The table name of the Azure SQL database. Type: string (or Expression with - resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureSqlTableDataset, self).__init__(**kwargs) - self.type = 'AzureSqlTable' # type: str - self.table_name = kwargs.get('table_name', None) - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - self.table = kwargs.get('table', None) - - -class AzureStorageLinkedService(LinkedService): - """The storage account linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: The connection string. It is mutually exclusive with sasUri property. - Type: string, SecureString or AzureKeyVaultSecretReference. - :type connection_string: any - :param account_key: The Azure key vault secret reference of accountKey in connection string. - :type account_key: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param sas_uri: SAS URI of the Azure Storage resource. It is mutually exclusive with - connectionString property. Type: string, SecureString or AzureKeyVaultSecretReference. - :type sas_uri: any - :param sas_token: The Azure key vault secret reference of sasToken in sas uri. - :type sas_token: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'account_key': {'key': 'typeProperties.accountKey', 'type': 'AzureKeyVaultSecretReference'}, - 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'object'}, - 'sas_token': {'key': 'typeProperties.sasToken', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureStorageLinkedService, self).__init__(**kwargs) - self.type = 'AzureStorage' # type: str - self.connection_string = kwargs.get('connection_string', None) - self.account_key = kwargs.get('account_key', None) - self.sas_uri = kwargs.get('sas_uri', None) - self.sas_token = kwargs.get('sas_token', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class AzureTableDataset(Dataset): - """The Azure Table storage dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: Required. The table name of the Azure Table storage. Type: string (or - Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'table_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureTableDataset, self).__init__(**kwargs) - self.type = 'AzureTable' # type: str - self.table_name = kwargs['table_name'] - - -class AzureTableSink(CopySink): - """A copy activity Azure Table sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param azure_table_default_partition_key_value: Azure Table default partition key value. Type: - string (or Expression with resultType string). - :type azure_table_default_partition_key_value: any - :param azure_table_partition_key_name: Azure Table partition key name. Type: string (or - Expression with resultType string). - :type azure_table_partition_key_name: any - :param azure_table_row_key_name: Azure Table row key name. Type: string (or Expression with - resultType string). - :type azure_table_row_key_name: any - :param azure_table_insert_type: Azure Table insert type. Type: string (or Expression with - resultType string). - :type azure_table_insert_type: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'azure_table_default_partition_key_value': {'key': 'azureTableDefaultPartitionKeyValue', 'type': 'object'}, - 'azure_table_partition_key_name': {'key': 'azureTablePartitionKeyName', 'type': 'object'}, - 'azure_table_row_key_name': {'key': 'azureTableRowKeyName', 'type': 'object'}, - 'azure_table_insert_type': {'key': 'azureTableInsertType', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureTableSink, self).__init__(**kwargs) - self.type = 'AzureTableSink' # type: str - self.azure_table_default_partition_key_value = kwargs.get('azure_table_default_partition_key_value', None) - self.azure_table_partition_key_name = kwargs.get('azure_table_partition_key_name', None) - self.azure_table_row_key_name = kwargs.get('azure_table_row_key_name', None) - self.azure_table_insert_type = kwargs.get('azure_table_insert_type', None) - - -class AzureTableSource(TabularSource): - """A copy activity Azure Table source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param azure_table_source_query: Azure Table source query. Type: string (or Expression with - resultType string). - :type azure_table_source_query: any - :param azure_table_source_ignore_table_not_found: Azure Table source ignore table not found. - Type: boolean (or Expression with resultType boolean). - :type azure_table_source_ignore_table_not_found: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'azure_table_source_query': {'key': 'azureTableSourceQuery', 'type': 'object'}, - 'azure_table_source_ignore_table_not_found': {'key': 'azureTableSourceIgnoreTableNotFound', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureTableSource, self).__init__(**kwargs) - self.type = 'AzureTableSource' # type: str - self.azure_table_source_query = kwargs.get('azure_table_source_query', None) - self.azure_table_source_ignore_table_not_found = kwargs.get('azure_table_source_ignore_table_not_found', None) - - -class AzureTableStorageLinkedService(LinkedService): - """The azure table storage linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: The connection string. It is mutually exclusive with sasUri property. - Type: string, SecureString or AzureKeyVaultSecretReference. - :type connection_string: any - :param account_key: The Azure key vault secret reference of accountKey in connection string. - :type account_key: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param sas_uri: SAS URI of the Azure Storage resource. It is mutually exclusive with - connectionString property. Type: string, SecureString or AzureKeyVaultSecretReference. - :type sas_uri: any - :param sas_token: The Azure key vault secret reference of sasToken in sas uri. - :type sas_token: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'account_key': {'key': 'typeProperties.accountKey', 'type': 'AzureKeyVaultSecretReference'}, - 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'object'}, - 'sas_token': {'key': 'typeProperties.sasToken', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureTableStorageLinkedService, self).__init__(**kwargs) - self.type = 'AzureTableStorage' # type: str - self.connection_string = kwargs.get('connection_string', None) - self.account_key = kwargs.get('account_key', None) - self.sas_uri = kwargs.get('sas_uri', None) - self.sas_token = kwargs.get('sas_token', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class BigDataPoolReference(msrest.serialization.Model): - """Big data pool reference. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Big data pool reference type. Possible values include: - "BigDataPoolReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.BigDataPoolReferenceType - :param reference_name: Required. Reference big data pool name. - :type reference_name: str - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(BigDataPoolReference, self).__init__(**kwargs) - self.type = kwargs['type'] - self.reference_name = kwargs['reference_name'] - - -class TrackedResource(Resource): - """The resource model definition for an Azure Resource Manager tracked top level resource which has 'tags' and a 'location'. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :param tags: A set of tags. Resource tags. - :type tags: dict[str, str] - :param location: Required. The geo-location where the resource lives. - :type location: str - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'location': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'tags': {'key': 'tags', 'type': '{str}'}, - 'location': {'key': 'location', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(TrackedResource, self).__init__(**kwargs) - self.tags = kwargs.get('tags', None) - self.location = kwargs['location'] - - -class BigDataPoolResourceInfo(TrackedResource): - """A Big Data pool. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :param tags: A set of tags. Resource tags. - :type tags: dict[str, str] - :param location: Required. The geo-location where the resource lives. - :type location: str - :param provisioning_state: The state of the Big Data pool. - :type provisioning_state: str - :param auto_scale: Auto-scaling properties. - :type auto_scale: ~azure.synapse.artifacts.v2020_12_01.models.AutoScaleProperties - :param creation_date: The time when the Big Data pool was created. - :type creation_date: ~datetime.datetime - :param auto_pause: Auto-pausing properties. - :type auto_pause: ~azure.synapse.artifacts.v2020_12_01.models.AutoPauseProperties - :param is_compute_isolation_enabled: Whether compute isolation is required or not. - :type is_compute_isolation_enabled: bool - :param session_level_packages_enabled: Whether session level packages enabled. - :type session_level_packages_enabled: bool - :param cache_size: The cache size. - :type cache_size: int - :param dynamic_executor_allocation: Dynamic Executor Allocation. - :type dynamic_executor_allocation: - ~azure.synapse.artifacts.v2020_12_01.models.DynamicExecutorAllocation - :param spark_events_folder: The Spark events folder. - :type spark_events_folder: str - :param node_count: The number of nodes in the Big Data pool. - :type node_count: int - :param library_requirements: Library version requirements. - :type library_requirements: ~azure.synapse.artifacts.v2020_12_01.models.LibraryRequirements - :param custom_libraries: List of custom libraries/packages associated with the spark pool. - :type custom_libraries: list[~azure.synapse.artifacts.v2020_12_01.models.LibraryInfo] - :param spark_config_properties: Spark configuration file to specify additional properties. - :type spark_config_properties: ~azure.synapse.artifacts.v2020_12_01.models.LibraryRequirements - :param spark_version: The Apache Spark version. - :type spark_version: str - :param default_spark_log_folder: The default folder where Spark logs will be written. - :type default_spark_log_folder: str - :param node_size: The level of compute power that each node in the Big Data pool has. Possible - values include: "None", "Small", "Medium", "Large", "XLarge", "XXLarge", "XXXLarge". - :type node_size: str or ~azure.synapse.artifacts.v2020_12_01.models.NodeSize - :param node_size_family: The kind of nodes that the Big Data pool provides. Possible values - include: "None", "MemoryOptimized". - :type node_size_family: str or ~azure.synapse.artifacts.v2020_12_01.models.NodeSizeFamily - :ivar last_succeeded_timestamp: The time when the Big Data pool was updated successfully. - :vartype last_succeeded_timestamp: ~datetime.datetime - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'location': {'required': True}, - 'last_succeeded_timestamp': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'tags': {'key': 'tags', 'type': '{str}'}, - 'location': {'key': 'location', 'type': 'str'}, - 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, - 'auto_scale': {'key': 'properties.autoScale', 'type': 'AutoScaleProperties'}, - 'creation_date': {'key': 'properties.creationDate', 'type': 'iso-8601'}, - 'auto_pause': {'key': 'properties.autoPause', 'type': 'AutoPauseProperties'}, - 'is_compute_isolation_enabled': {'key': 'properties.isComputeIsolationEnabled', 'type': 'bool'}, - 'session_level_packages_enabled': {'key': 'properties.sessionLevelPackagesEnabled', 'type': 'bool'}, - 'cache_size': {'key': 'properties.cacheSize', 'type': 'int'}, - 'dynamic_executor_allocation': {'key': 'properties.dynamicExecutorAllocation', 'type': 'DynamicExecutorAllocation'}, - 'spark_events_folder': {'key': 'properties.sparkEventsFolder', 'type': 'str'}, - 'node_count': {'key': 'properties.nodeCount', 'type': 'int'}, - 'library_requirements': {'key': 'properties.libraryRequirements', 'type': 'LibraryRequirements'}, - 'custom_libraries': {'key': 'properties.customLibraries', 'type': '[LibraryInfo]'}, - 'spark_config_properties': {'key': 'properties.sparkConfigProperties', 'type': 'LibraryRequirements'}, - 'spark_version': {'key': 'properties.sparkVersion', 'type': 'str'}, - 'default_spark_log_folder': {'key': 'properties.defaultSparkLogFolder', 'type': 'str'}, - 'node_size': {'key': 'properties.nodeSize', 'type': 'str'}, - 'node_size_family': {'key': 'properties.nodeSizeFamily', 'type': 'str'}, - 'last_succeeded_timestamp': {'key': 'properties.lastSucceededTimestamp', 'type': 'iso-8601'}, - } - - def __init__( - self, - **kwargs - ): - super(BigDataPoolResourceInfo, self).__init__(**kwargs) - self.provisioning_state = kwargs.get('provisioning_state', None) - self.auto_scale = kwargs.get('auto_scale', None) - self.creation_date = kwargs.get('creation_date', None) - self.auto_pause = kwargs.get('auto_pause', None) - self.is_compute_isolation_enabled = kwargs.get('is_compute_isolation_enabled', None) - self.session_level_packages_enabled = kwargs.get('session_level_packages_enabled', None) - self.cache_size = kwargs.get('cache_size', None) - self.dynamic_executor_allocation = kwargs.get('dynamic_executor_allocation', None) - self.spark_events_folder = kwargs.get('spark_events_folder', None) - self.node_count = kwargs.get('node_count', None) - self.library_requirements = kwargs.get('library_requirements', None) - self.custom_libraries = kwargs.get('custom_libraries', None) - self.spark_config_properties = kwargs.get('spark_config_properties', None) - self.spark_version = kwargs.get('spark_version', None) - self.default_spark_log_folder = kwargs.get('default_spark_log_folder', None) - self.node_size = kwargs.get('node_size', None) - self.node_size_family = kwargs.get('node_size_family', None) - self.last_succeeded_timestamp = None - - -class BigDataPoolResourceInfoListResult(msrest.serialization.Model): - """Collection of Big Data pool information. - - :param next_link: Link to the next page of results. - :type next_link: str - :param value: List of Big Data pools. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.BigDataPoolResourceInfo] - """ - - _attribute_map = { - 'next_link': {'key': 'nextLink', 'type': 'str'}, - 'value': {'key': 'value', 'type': '[BigDataPoolResourceInfo]'}, - } - - def __init__( - self, - **kwargs - ): - super(BigDataPoolResourceInfoListResult, self).__init__(**kwargs) - self.next_link = kwargs.get('next_link', None) - self.value = kwargs.get('value', None) - - -class BinaryDataset(Dataset): - """Binary dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param location: The location of the Binary storage. - :type location: ~azure.synapse.artifacts.v2020_12_01.models.DatasetLocation - :param compression: The data compression method used for the binary dataset. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'location': {'key': 'typeProperties.location', 'type': 'DatasetLocation'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - } - - def __init__( - self, - **kwargs - ): - super(BinaryDataset, self).__init__(**kwargs) - self.type = 'Binary' # type: str - self.location = kwargs.get('location', None) - self.compression = kwargs.get('compression', None) - - -class FormatReadSettings(msrest.serialization.Model): - """Format read settings. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: BinaryReadSettings, DelimitedTextReadSettings, JsonReadSettings, XmlReadSettings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - _subtype_map = { - 'type': {'BinaryReadSettings': 'BinaryReadSettings', 'DelimitedTextReadSettings': 'DelimitedTextReadSettings', 'JsonReadSettings': 'JsonReadSettings', 'XmlReadSettings': 'XmlReadSettings'} - } - - def __init__( - self, - **kwargs - ): - super(FormatReadSettings, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = 'FormatReadSettings' # type: str - - -class BinaryReadSettings(FormatReadSettings): - """Binary read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param compression_properties: Compression settings. - :type compression_properties: - ~azure.synapse.artifacts.v2020_12_01.models.CompressionReadSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'compression_properties': {'key': 'compressionProperties', 'type': 'CompressionReadSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(BinaryReadSettings, self).__init__(**kwargs) - self.type = 'BinaryReadSettings' # type: str - self.compression_properties = kwargs.get('compression_properties', None) - - -class BinarySink(CopySink): - """A copy activity Binary sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Binary store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreWriteSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreWriteSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(BinarySink, self).__init__(**kwargs) - self.type = 'BinarySink' # type: str - self.store_settings = kwargs.get('store_settings', None) - - -class BinarySource(CopySource): - """A copy activity Binary source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Binary store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - :param format_settings: Binary format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.BinaryReadSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'format_settings': {'key': 'formatSettings', 'type': 'BinaryReadSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(BinarySource, self).__init__(**kwargs) - self.type = 'BinarySource' # type: str - self.store_settings = kwargs.get('store_settings', None) - self.format_settings = kwargs.get('format_settings', None) - - -class Trigger(msrest.serialization.Model): - """Azure Synapse nested object which contains information about creating pipeline run. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: ChainingTrigger, MultiplePipelineTrigger, RerunTumblingWindowTrigger, TumblingWindowTrigger. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Trigger type.Constant filled by server. - :type type: str - :param description: Trigger description. - :type description: str - :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are - called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRuntimeState - :param annotations: List of tags that can be used for describing the trigger. - :type annotations: list[any] - """ - - _validation = { - 'type': {'required': True}, - 'runtime_state': {'readonly': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - } - - _subtype_map = { - 'type': {'ChainingTrigger': 'ChainingTrigger', 'MultiplePipelineTrigger': 'MultiplePipelineTrigger', 'RerunTumblingWindowTrigger': 'RerunTumblingWindowTrigger', 'TumblingWindowTrigger': 'TumblingWindowTrigger'} - } - - def __init__( - self, - **kwargs - ): - super(Trigger, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = 'Trigger' # type: str - self.description = kwargs.get('description', None) - self.runtime_state = None - self.annotations = kwargs.get('annotations', None) - - -class MultiplePipelineTrigger(Trigger): - """Base class for all triggers that support one to many model for trigger to pipeline. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: BlobEventsTrigger, BlobTrigger, CustomEventsTrigger, ScheduleTrigger. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Trigger type.Constant filled by server. - :type type: str - :param description: Trigger description. - :type description: str - :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are - called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRuntimeState - :param annotations: List of tags that can be used for describing the trigger. - :type annotations: list[any] - :param pipelines: Pipelines that need to be started. - :type pipelines: list[~azure.synapse.artifacts.v2020_12_01.models.TriggerPipelineReference] - """ - - _validation = { - 'type': {'required': True}, - 'runtime_state': {'readonly': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, - } - - _subtype_map = { - 'type': {'BlobEventsTrigger': 'BlobEventsTrigger', 'BlobTrigger': 'BlobTrigger', 'CustomEventsTrigger': 'CustomEventsTrigger', 'ScheduleTrigger': 'ScheduleTrigger'} - } - - def __init__( - self, - **kwargs - ): - super(MultiplePipelineTrigger, self).__init__(**kwargs) - self.type = 'MultiplePipelineTrigger' # type: str - self.pipelines = kwargs.get('pipelines', None) - - -class BlobEventsTrigger(MultiplePipelineTrigger): - """Trigger that runs every time a Blob event occurs. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Trigger type.Constant filled by server. - :type type: str - :param description: Trigger description. - :type description: str - :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are - called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRuntimeState - :param annotations: List of tags that can be used for describing the trigger. - :type annotations: list[any] - :param pipelines: Pipelines that need to be started. - :type pipelines: list[~azure.synapse.artifacts.v2020_12_01.models.TriggerPipelineReference] - :param blob_path_begins_with: The blob path must begin with the pattern provided for trigger to - fire. For example, '/records/blobs/december/' will only fire the trigger for blobs in the - december folder under the records container. At least one of these must be provided: - blobPathBeginsWith, blobPathEndsWith. - :type blob_path_begins_with: str - :param blob_path_ends_with: The blob path must end with the pattern provided for trigger to - fire. For example, 'december/boxes.csv' will only fire the trigger for blobs named boxes in a - december folder. At least one of these must be provided: blobPathBeginsWith, blobPathEndsWith. - :type blob_path_ends_with: str - :param ignore_empty_blobs: If set to true, blobs with zero bytes will be ignored. - :type ignore_empty_blobs: bool - :param events: Required. The type of events that cause this trigger to fire. - :type events: list[str or ~azure.synapse.artifacts.v2020_12_01.models.BlobEventType] - :param scope: Required. The ARM resource ID of the Storage Account. - :type scope: str - """ - - _validation = { - 'type': {'required': True}, - 'runtime_state': {'readonly': True}, - 'events': {'required': True}, - 'scope': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, - 'blob_path_begins_with': {'key': 'typeProperties.blobPathBeginsWith', 'type': 'str'}, - 'blob_path_ends_with': {'key': 'typeProperties.blobPathEndsWith', 'type': 'str'}, - 'ignore_empty_blobs': {'key': 'typeProperties.ignoreEmptyBlobs', 'type': 'bool'}, - 'events': {'key': 'typeProperties.events', 'type': '[str]'}, - 'scope': {'key': 'typeProperties.scope', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(BlobEventsTrigger, self).__init__(**kwargs) - self.type = 'BlobEventsTrigger' # type: str - self.blob_path_begins_with = kwargs.get('blob_path_begins_with', None) - self.blob_path_ends_with = kwargs.get('blob_path_ends_with', None) - self.ignore_empty_blobs = kwargs.get('ignore_empty_blobs', None) - self.events = kwargs['events'] - self.scope = kwargs['scope'] - - -class BlobSink(CopySink): - """A copy activity Azure Blob sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param blob_writer_overwrite_files: Blob writer overwrite files. Type: boolean (or Expression - with resultType boolean). - :type blob_writer_overwrite_files: any - :param blob_writer_date_time_format: Blob writer date time format. Type: string (or Expression - with resultType string). - :type blob_writer_date_time_format: any - :param blob_writer_add_header: Blob writer add header. Type: boolean (or Expression with - resultType boolean). - :type blob_writer_add_header: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'blob_writer_overwrite_files': {'key': 'blobWriterOverwriteFiles', 'type': 'object'}, - 'blob_writer_date_time_format': {'key': 'blobWriterDateTimeFormat', 'type': 'object'}, - 'blob_writer_add_header': {'key': 'blobWriterAddHeader', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(BlobSink, self).__init__(**kwargs) - self.type = 'BlobSink' # type: str - self.blob_writer_overwrite_files = kwargs.get('blob_writer_overwrite_files', None) - self.blob_writer_date_time_format = kwargs.get('blob_writer_date_time_format', None) - self.blob_writer_add_header = kwargs.get('blob_writer_add_header', None) - self.copy_behavior = kwargs.get('copy_behavior', None) - - -class BlobSource(CopySource): - """A copy activity Azure Blob source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param treat_empty_as_null: Treat empty as null. Type: boolean (or Expression with resultType - boolean). - :type treat_empty_as_null: any - :param skip_header_line_count: Number of header lines to skip from each blob. Type: integer (or - Expression with resultType integer). - :type skip_header_line_count: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'treat_empty_as_null': {'key': 'treatEmptyAsNull', 'type': 'object'}, - 'skip_header_line_count': {'key': 'skipHeaderLineCount', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(BlobSource, self).__init__(**kwargs) - self.type = 'BlobSource' # type: str - self.treat_empty_as_null = kwargs.get('treat_empty_as_null', None) - self.skip_header_line_count = kwargs.get('skip_header_line_count', None) - self.recursive = kwargs.get('recursive', None) - - -class BlobTrigger(MultiplePipelineTrigger): - """Trigger that runs every time the selected Blob container changes. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Trigger type.Constant filled by server. - :type type: str - :param description: Trigger description. - :type description: str - :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are - called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRuntimeState - :param annotations: List of tags that can be used for describing the trigger. - :type annotations: list[any] - :param pipelines: Pipelines that need to be started. - :type pipelines: list[~azure.synapse.artifacts.v2020_12_01.models.TriggerPipelineReference] - :param folder_path: Required. The path of the container/folder that will trigger the pipeline. - :type folder_path: str - :param max_concurrency: Required. The max number of parallel files to handle when it is - triggered. - :type max_concurrency: int - :param linked_service: Required. The Azure Storage linked service reference. - :type linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - """ - - _validation = { - 'type': {'required': True}, - 'runtime_state': {'readonly': True}, - 'folder_path': {'required': True}, - 'max_concurrency': {'required': True}, - 'linked_service': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, - 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'str'}, - 'max_concurrency': {'key': 'typeProperties.maxConcurrency', 'type': 'int'}, - 'linked_service': {'key': 'typeProperties.linkedService', 'type': 'LinkedServiceReference'}, - } - - def __init__( - self, - **kwargs - ): - super(BlobTrigger, self).__init__(**kwargs) - self.type = 'BlobTrigger' # type: str - self.folder_path = kwargs['folder_path'] - self.max_concurrency = kwargs['max_concurrency'] - self.linked_service = kwargs['linked_service'] - - -class CassandraLinkedService(LinkedService): - """Linked service for Cassandra data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. Host name for connection. Type: string (or Expression with resultType - string). - :type host: any - :param authentication_type: AuthenticationType to be used for connection. Type: string (or - Expression with resultType string). - :type authentication_type: any - :param port: The port for the connection. Type: integer (or Expression with resultType - integer). - :type port: any - :param username: Username for authentication. Type: string (or Expression with resultType - string). - :type username: any - :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CassandraLinkedService, self).__init__(**kwargs) - self.type = 'Cassandra' # type: str - self.host = kwargs['host'] - self.authentication_type = kwargs.get('authentication_type', None) - self.port = kwargs.get('port', None) - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class CassandraSource(TabularSource): - """A copy activity source for a Cassandra database. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Should be a SQL-92 query expression or Cassandra Query Language - (CQL) command. Type: string (or Expression with resultType string). - :type query: any - :param consistency_level: The consistency level specifies how many Cassandra servers must - respond to a read request before returning data to the client application. Cassandra checks the - specified number of Cassandra servers for data to satisfy the read request. Must be one of - cassandraSourceReadConsistencyLevels. The default value is 'ONE'. It is case-insensitive. - Possible values include: "ALL", "EACH_QUORUM", "QUORUM", "LOCAL_QUORUM", "ONE", "TWO", "THREE", - "LOCAL_ONE", "SERIAL", "LOCAL_SERIAL". - :type consistency_level: str or - ~azure.synapse.artifacts.v2020_12_01.models.CassandraSourceReadConsistencyLevels - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'consistency_level': {'key': 'consistencyLevel', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(CassandraSource, self).__init__(**kwargs) - self.type = 'CassandraSource' # type: str - self.query = kwargs.get('query', None) - self.consistency_level = kwargs.get('consistency_level', None) - - -class CassandraTableDataset(Dataset): - """The Cassandra database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name of the Cassandra database. Type: string (or Expression with - resultType string). - :type table_name: any - :param keyspace: The keyspace of the Cassandra database. Type: string (or Expression with - resultType string). - :type keyspace: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'keyspace': {'key': 'typeProperties.keyspace', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CassandraTableDataset, self).__init__(**kwargs) - self.type = 'CassandraTable' # type: str - self.table_name = kwargs.get('table_name', None) - self.keyspace = kwargs.get('keyspace', None) - - -class ChainingTrigger(Trigger): - """Trigger that allows the referenced pipeline to depend on other pipeline runs based on runDimension Name/Value pairs. Upstream pipelines should declare the same runDimension Name and their runs should have the values for those runDimensions. The referenced pipeline run would be triggered if the values for the runDimension match for all upstream pipeline runs. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Trigger type.Constant filled by server. - :type type: str - :param description: Trigger description. - :type description: str - :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are - called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRuntimeState - :param annotations: List of tags that can be used for describing the trigger. - :type annotations: list[any] - :param pipeline: Required. Pipeline for which runs are created when all upstream pipelines - complete successfully. - :type pipeline: ~azure.synapse.artifacts.v2020_12_01.models.TriggerPipelineReference - :param depends_on: Required. Upstream Pipelines. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.PipelineReference] - :param run_dimension: Required. Run Dimension property that needs to be emitted by upstream - pipelines. - :type run_dimension: str - """ - - _validation = { - 'type': {'required': True}, - 'runtime_state': {'readonly': True}, - 'pipeline': {'required': True}, - 'depends_on': {'required': True}, - 'run_dimension': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'pipeline': {'key': 'pipeline', 'type': 'TriggerPipelineReference'}, - 'depends_on': {'key': 'typeProperties.dependsOn', 'type': '[PipelineReference]'}, - 'run_dimension': {'key': 'typeProperties.runDimension', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(ChainingTrigger, self).__init__(**kwargs) - self.type = 'ChainingTrigger' # type: str - self.pipeline = kwargs['pipeline'] - self.depends_on = kwargs['depends_on'] - self.run_dimension = kwargs['run_dimension'] - - -class CloudError(msrest.serialization.Model): - """The object that defines the structure of an Azure Synapse error response. - - All required parameters must be populated in order to send to Azure. - - :param code: Required. Error code. - :type code: str - :param message: Required. Error message. - :type message: str - :param target: Property name/path in request associated with error. - :type target: str - :param details: Array with additional error details. - :type details: list[~azure.synapse.artifacts.v2020_12_01.models.CloudError] - """ - - _validation = { - 'code': {'required': True}, - 'message': {'required': True}, - } - - _attribute_map = { - 'code': {'key': 'error.code', 'type': 'str'}, - 'message': {'key': 'error.message', 'type': 'str'}, - 'target': {'key': 'error.target', 'type': 'str'}, - 'details': {'key': 'error.details', 'type': '[CloudError]'}, - } - - def __init__( - self, - **kwargs - ): - super(CloudError, self).__init__(**kwargs) - self.code = kwargs['code'] - self.message = kwargs['message'] - self.target = kwargs.get('target', None) - self.details = kwargs.get('details', None) - - -class CommonDataServiceForAppsEntityDataset(Dataset): - """The Common Data Service for Apps entity dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param entity_name: The logical name of the entity. Type: string (or Expression with resultType - string). - :type entity_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'entity_name': {'key': 'typeProperties.entityName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CommonDataServiceForAppsEntityDataset, self).__init__(**kwargs) - self.type = 'CommonDataServiceForAppsEntity' # type: str - self.entity_name = kwargs.get('entity_name', None) - - -class CommonDataServiceForAppsLinkedService(LinkedService): - """Common Data Service for Apps linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param deployment_type: Required. The deployment type of the Common Data Service for Apps - instance. 'Online' for Common Data Service for Apps Online and 'OnPremisesWithIfd' for Common - Data Service for Apps on-premises with Ifd. Type: string (or Expression with resultType - string). Possible values include: "Online", "OnPremisesWithIfd". - :type deployment_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsDeploymentType - :param host_name: The host name of the on-premises Common Data Service for Apps server. The - property is required for on-prem and not allowed for online. Type: string (or Expression with - resultType string). - :type host_name: any - :param port: The port of on-premises Common Data Service for Apps server. The property is - required for on-prem and not allowed for online. Default is 443. Type: integer (or Expression - with resultType integer), minimum: 0. - :type port: any - :param service_uri: The URL to the Microsoft Common Data Service for Apps server. The property - is required for on-line and not allowed for on-prem. Type: string (or Expression with - resultType string). - :type service_uri: any - :param organization_name: The organization name of the Common Data Service for Apps instance. - The property is required for on-prem and required for online when there are more than one - Common Data Service for Apps instances associated with the user. Type: string (or Expression - with resultType string). - :type organization_name: any - :param authentication_type: Required. The authentication type to connect to Common Data Service - for Apps server. 'Office365' for online scenario, 'Ifd' for on-premises with Ifd scenario. - 'AADServicePrincipal' for Server-To-Server authentication in online scenario. Type: string (or - Expression with resultType string). Possible values include: "Office365", "Ifd", - "AADServicePrincipal". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsAuthenticationType - :param username: User name to access the Common Data Service for Apps instance. Type: string - (or Expression with resultType string). - :type username: any - :param password: Password to access the Common Data Service for Apps instance. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param service_principal_id: The client ID of the application in Azure Active Directory used - for Server-To-Server authentication. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_credential_type: A string from ServicePrincipalCredentialEnum or an - expression. - :type service_principal_credential_type: any - :param service_principal_credential: The credential of the service principal object in Azure - Active Directory. If servicePrincipalCredentialType is 'ServicePrincipalKey', - servicePrincipalCredential can be SecureString or AzureKeyVaultSecretReference. If - servicePrincipalCredentialType is 'ServicePrincipalCert', servicePrincipalCredential can only - be AzureKeyVaultSecretReference. - :type service_principal_credential: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'deployment_type': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'str'}, - 'host_name': {'key': 'typeProperties.hostName', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'service_uri': {'key': 'typeProperties.serviceUri', 'type': 'object'}, - 'organization_name': {'key': 'typeProperties.organizationName', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_credential_type': {'key': 'typeProperties.servicePrincipalCredentialType', 'type': 'object'}, - 'service_principal_credential': {'key': 'typeProperties.servicePrincipalCredential', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CommonDataServiceForAppsLinkedService, self).__init__(**kwargs) - self.type = 'CommonDataServiceForApps' # type: str - self.deployment_type = kwargs['deployment_type'] - self.host_name = kwargs.get('host_name', None) - self.port = kwargs.get('port', None) - self.service_uri = kwargs.get('service_uri', None) - self.organization_name = kwargs.get('organization_name', None) - self.authentication_type = kwargs['authentication_type'] - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.service_principal_id = kwargs.get('service_principal_id', None) - self.service_principal_credential_type = kwargs.get('service_principal_credential_type', None) - self.service_principal_credential = kwargs.get('service_principal_credential', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class CommonDataServiceForAppsSink(CopySink): - """A copy activity Common Data Service for Apps sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param write_behavior: Required. The write behavior for the operation. Possible values include: - "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsSinkWriteBehavior - :param ignore_null_values: The flag indicating whether to ignore null values from input dataset - (except key fields) during write operation. Default is false. Type: boolean (or Expression with - resultType boolean). - :type ignore_null_values: any - :param alternate_key_name: The logical name of the alternate key which will be used when - upserting records. Type: string (or Expression with resultType string). - :type alternate_key_name: any - """ - - _validation = { - 'type': {'required': True}, - 'write_behavior': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, - 'ignore_null_values': {'key': 'ignoreNullValues', 'type': 'object'}, - 'alternate_key_name': {'key': 'alternateKeyName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CommonDataServiceForAppsSink, self).__init__(**kwargs) - self.type = 'CommonDataServiceForAppsSink' # type: str - self.write_behavior = kwargs['write_behavior'] - self.ignore_null_values = kwargs.get('ignore_null_values', None) - self.alternate_key_name = kwargs.get('alternate_key_name', None) - - -class CommonDataServiceForAppsSource(CopySource): - """A copy activity Common Data Service for Apps source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: FetchXML is a proprietary query language that is used in Microsoft Common Data - Service for Apps (online & on-premises). Type: string (or Expression with resultType string). - :type query: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CommonDataServiceForAppsSource, self).__init__(**kwargs) - self.type = 'CommonDataServiceForAppsSource' # type: str - self.query = kwargs.get('query', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class CompressionReadSettings(msrest.serialization.Model): - """Compression read settings. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: TarGZipReadSettings, TarReadSettings, ZipDeflateReadSettings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The Compression setting type.Constant filled by server. - :type type: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - _subtype_map = { - 'type': {'TarGZipReadSettings': 'TarGZipReadSettings', 'TarReadSettings': 'TarReadSettings', 'ZipDeflateReadSettings': 'ZipDeflateReadSettings'} - } - - def __init__( - self, - **kwargs - ): - super(CompressionReadSettings, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = 'CompressionReadSettings' # type: str - - -class ConcurLinkedService(LinkedService): - """Concur Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_properties: Properties used to connect to Concur. It is mutually exclusive - with any other properties in the linked service. Type: object. - :type connection_properties: any - :param client_id: Required. Application client_id supplied by Concur App Management. - :type client_id: any - :param username: Required. The user name that you use to access Concur Service. - :type username: any - :param password: The password corresponding to the user name that you provided in the username - field. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'client_id': {'required': True}, - 'username': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_properties': {'key': 'typeProperties.connectionProperties', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ConcurLinkedService, self).__init__(**kwargs) - self.type = 'Concur' # type: str - self.connection_properties = kwargs.get('connection_properties', None) - self.client_id = kwargs['client_id'] - self.username = kwargs['username'] - self.password = kwargs.get('password', None) - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.use_host_verification = kwargs.get('use_host_verification', None) - self.use_peer_verification = kwargs.get('use_peer_verification', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class ConcurObjectDataset(Dataset): - """Concur Service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ConcurObjectDataset, self).__init__(**kwargs) - self.type = 'ConcurObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class ConcurSource(TabularSource): - """A copy activity Concur Service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ConcurSource, self).__init__(**kwargs) - self.type = 'ConcurSource' # type: str - self.query = kwargs.get('query', None) - - -class CopyActivity(ExecutionActivity): - """Copy activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param inputs: List of inputs for the activity. - :type inputs: list[~azure.synapse.artifacts.v2020_12_01.models.DatasetReference] - :param outputs: List of outputs for the activity. - :type outputs: list[~azure.synapse.artifacts.v2020_12_01.models.DatasetReference] - :param source: Required. Copy activity source. - :type source: ~azure.synapse.artifacts.v2020_12_01.models.CopySource - :param sink: Required. Copy activity sink. - :type sink: ~azure.synapse.artifacts.v2020_12_01.models.CopySink - :param translator: Copy activity translator. If not specified, tabular translator is used. - :type translator: any - :param enable_staging: Specifies whether to copy data via an interim staging. Default value is - false. Type: boolean (or Expression with resultType boolean). - :type enable_staging: any - :param staging_settings: Specifies interim staging settings when EnableStaging is true. - :type staging_settings: ~azure.synapse.artifacts.v2020_12_01.models.StagingSettings - :param parallel_copies: Maximum number of concurrent sessions opened on the source or sink to - avoid overloading the data store. Type: integer (or Expression with resultType integer), - minimum: 0. - :type parallel_copies: any - :param data_integration_units: Maximum number of data integration units that can be used to - perform this data movement. Type: integer (or Expression with resultType integer), minimum: 0. - :type data_integration_units: any - :param enable_skip_incompatible_row: Whether to skip incompatible row. Default value is false. - Type: boolean (or Expression with resultType boolean). - :type enable_skip_incompatible_row: any - :param redirect_incompatible_row_settings: Redirect incompatible row settings when - EnableSkipIncompatibleRow is true. - :type redirect_incompatible_row_settings: - ~azure.synapse.artifacts.v2020_12_01.models.RedirectIncompatibleRowSettings - :param log_storage_settings: (Deprecated. Please use LogSettings) Log storage settings customer - need to provide when enabling session log. - :type log_storage_settings: ~azure.synapse.artifacts.v2020_12_01.models.LogStorageSettings - :param log_settings: Log settings customer needs provide when enabling log. - :type log_settings: ~azure.synapse.artifacts.v2020_12_01.models.LogSettings - :param preserve_rules: Preserve Rules. - :type preserve_rules: list[any] - :param preserve: Preserve rules. - :type preserve: list[any] - :param validate_data_consistency: Whether to enable Data Consistency validation. Type: boolean - (or Expression with resultType boolean). - :type validate_data_consistency: any - :param skip_error_file: Specify the fault tolerance for data consistency. - :type skip_error_file: ~azure.synapse.artifacts.v2020_12_01.models.SkipErrorFile - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'source': {'required': True}, - 'sink': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'inputs': {'key': 'inputs', 'type': '[DatasetReference]'}, - 'outputs': {'key': 'outputs', 'type': '[DatasetReference]'}, - 'source': {'key': 'typeProperties.source', 'type': 'CopySource'}, - 'sink': {'key': 'typeProperties.sink', 'type': 'CopySink'}, - 'translator': {'key': 'typeProperties.translator', 'type': 'object'}, - 'enable_staging': {'key': 'typeProperties.enableStaging', 'type': 'object'}, - 'staging_settings': {'key': 'typeProperties.stagingSettings', 'type': 'StagingSettings'}, - 'parallel_copies': {'key': 'typeProperties.parallelCopies', 'type': 'object'}, - 'data_integration_units': {'key': 'typeProperties.dataIntegrationUnits', 'type': 'object'}, - 'enable_skip_incompatible_row': {'key': 'typeProperties.enableSkipIncompatibleRow', 'type': 'object'}, - 'redirect_incompatible_row_settings': {'key': 'typeProperties.redirectIncompatibleRowSettings', 'type': 'RedirectIncompatibleRowSettings'}, - 'log_storage_settings': {'key': 'typeProperties.logStorageSettings', 'type': 'LogStorageSettings'}, - 'log_settings': {'key': 'typeProperties.logSettings', 'type': 'LogSettings'}, - 'preserve_rules': {'key': 'typeProperties.preserveRules', 'type': '[object]'}, - 'preserve': {'key': 'typeProperties.preserve', 'type': '[object]'}, - 'validate_data_consistency': {'key': 'typeProperties.validateDataConsistency', 'type': 'object'}, - 'skip_error_file': {'key': 'typeProperties.skipErrorFile', 'type': 'SkipErrorFile'}, - } - - def __init__( - self, - **kwargs - ): - super(CopyActivity, self).__init__(**kwargs) - self.type = 'Copy' # type: str - self.inputs = kwargs.get('inputs', None) - self.outputs = kwargs.get('outputs', None) - self.source = kwargs['source'] - self.sink = kwargs['sink'] - self.translator = kwargs.get('translator', None) - self.enable_staging = kwargs.get('enable_staging', None) - self.staging_settings = kwargs.get('staging_settings', None) - self.parallel_copies = kwargs.get('parallel_copies', None) - self.data_integration_units = kwargs.get('data_integration_units', None) - self.enable_skip_incompatible_row = kwargs.get('enable_skip_incompatible_row', None) - self.redirect_incompatible_row_settings = kwargs.get('redirect_incompatible_row_settings', None) - self.log_storage_settings = kwargs.get('log_storage_settings', None) - self.log_settings = kwargs.get('log_settings', None) - self.preserve_rules = kwargs.get('preserve_rules', None) - self.preserve = kwargs.get('preserve', None) - self.validate_data_consistency = kwargs.get('validate_data_consistency', None) - self.skip_error_file = kwargs.get('skip_error_file', None) - - -class CopyActivityLogSettings(msrest.serialization.Model): - """Settings for copy activity log. - - :param log_level: Gets or sets the log level, support: Info, Warning. Type: string (or - Expression with resultType string). - :type log_level: any - :param enable_reliable_logging: Specifies whether to enable reliable logging. Type: boolean (or - Expression with resultType boolean). - :type enable_reliable_logging: any - """ - - _attribute_map = { - 'log_level': {'key': 'logLevel', 'type': 'object'}, - 'enable_reliable_logging': {'key': 'enableReliableLogging', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CopyActivityLogSettings, self).__init__(**kwargs) - self.log_level = kwargs.get('log_level', None) - self.enable_reliable_logging = kwargs.get('enable_reliable_logging', None) - - -class CopyTranslator(msrest.serialization.Model): - """A copy activity translator. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: TabularTranslator. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy translator type.Constant filled by server. - :type type: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - _subtype_map = { - 'type': {'TabularTranslator': 'TabularTranslator'} - } - - def __init__( - self, - **kwargs - ): - super(CopyTranslator, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = 'CopyTranslator' # type: str - - -class CosmosDbLinkedService(LinkedService): - """Microsoft Azure Cosmos Database (CosmosDB) linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: The connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param account_endpoint: The endpoint of the Azure CosmosDB account. Type: string (or - Expression with resultType string). - :type account_endpoint: any - :param database: The name of the database. Type: string (or Expression with resultType string). - :type database: any - :param account_key: The account key of the Azure CosmosDB account. Type: SecureString or - AzureKeyVaultSecretReference. - :type account_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'account_endpoint': {'key': 'typeProperties.accountEndpoint', 'type': 'object'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'account_key': {'key': 'typeProperties.accountKey', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CosmosDbLinkedService, self).__init__(**kwargs) - self.type = 'CosmosDb' # type: str - self.connection_string = kwargs.get('connection_string', None) - self.account_endpoint = kwargs.get('account_endpoint', None) - self.database = kwargs.get('database', None) - self.account_key = kwargs.get('account_key', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class CosmosDbMongoDbApiCollectionDataset(Dataset): - """The CosmosDB (MongoDB API) database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param collection: Required. The collection name of the CosmosDB (MongoDB API) database. Type: - string (or Expression with resultType string). - :type collection: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'collection': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'collection': {'key': 'typeProperties.collection', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CosmosDbMongoDbApiCollectionDataset, self).__init__(**kwargs) - self.type = 'CosmosDbMongoDbApiCollection' # type: str - self.collection = kwargs['collection'] - - -class CosmosDbMongoDbApiLinkedService(LinkedService): - """Linked service for CosmosDB (MongoDB API) data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The CosmosDB (MongoDB API) connection string. Type: string, - SecureString or AzureKeyVaultSecretReference. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param database: Required. The name of the CosmosDB (MongoDB API) database that you want to - access. Type: string (or Expression with resultType string). - :type database: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - 'database': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CosmosDbMongoDbApiLinkedService, self).__init__(**kwargs) - self.type = 'CosmosDbMongoDbApi' # type: str - self.connection_string = kwargs['connection_string'] - self.database = kwargs['database'] - - -class CosmosDbMongoDbApiSink(CopySink): - """A copy activity sink for a CosmosDB (MongoDB API) database. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param write_behavior: Specifies whether the document with same key to be overwritten (upsert) - rather than throw exception (insert). The default value is "insert". Type: string (or - Expression with resultType string). Type: string (or Expression with resultType string). - :type write_behavior: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CosmosDbMongoDbApiSink, self).__init__(**kwargs) - self.type = 'CosmosDbMongoDbApiSink' # type: str - self.write_behavior = kwargs.get('write_behavior', None) - - -class CosmosDbMongoDbApiSource(CopySource): - """A copy activity source for a CosmosDB (MongoDB API) database. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param filter: Specifies selection filter using query operators. To return all documents in a - collection, omit this parameter or pass an empty document ({}). Type: string (or Expression - with resultType string). - :type filter: any - :param cursor_methods: Cursor methods for Mongodb query. - :type cursor_methods: - ~azure.synapse.artifacts.v2020_12_01.models.MongoDbCursorMethodsProperties - :param batch_size: Specifies the number of documents to return in each batch of the response - from MongoDB instance. In most cases, modifying the batch size will not affect the user or the - application. This property's main purpose is to avoid hit the limitation of response size. - Type: integer (or Expression with resultType integer). - :type batch_size: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'filter': {'key': 'filter', 'type': 'object'}, - 'cursor_methods': {'key': 'cursorMethods', 'type': 'MongoDbCursorMethodsProperties'}, - 'batch_size': {'key': 'batchSize', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CosmosDbMongoDbApiSource, self).__init__(**kwargs) - self.type = 'CosmosDbMongoDbApiSource' # type: str - self.filter = kwargs.get('filter', None) - self.cursor_methods = kwargs.get('cursor_methods', None) - self.batch_size = kwargs.get('batch_size', None) - self.query_timeout = kwargs.get('query_timeout', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class CosmosDbSqlApiCollectionDataset(Dataset): - """Microsoft Azure CosmosDB (SQL API) Collection dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param collection_name: Required. CosmosDB (SQL API) collection name. Type: string (or - Expression with resultType string). - :type collection_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'collection_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'collection_name': {'key': 'typeProperties.collectionName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CosmosDbSqlApiCollectionDataset, self).__init__(**kwargs) - self.type = 'CosmosDbSqlApiCollection' # type: str - self.collection_name = kwargs['collection_name'] - - -class CosmosDbSqlApiSink(CopySink): - """A copy activity Azure CosmosDB (SQL API) Collection sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param write_behavior: Describes how to write data to Azure Cosmos DB. Type: string (or - Expression with resultType string). Allowed values: insert and upsert. - :type write_behavior: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CosmosDbSqlApiSink, self).__init__(**kwargs) - self.type = 'CosmosDbSqlApiSink' # type: str - self.write_behavior = kwargs.get('write_behavior', None) - - -class CosmosDbSqlApiSource(CopySource): - """A copy activity Azure CosmosDB (SQL API) Collection source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: SQL API query. Type: string (or Expression with resultType string). - :type query: any - :param page_size: Page size of the result. Type: integer (or Expression with resultType - integer). - :type page_size: any - :param preferred_regions: Preferred regions. Type: array of strings (or Expression with - resultType array of strings). - :type preferred_regions: any - :param detect_datetime: Whether detect primitive values as datetime values. Type: boolean (or - Expression with resultType boolean). - :type detect_datetime: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'page_size': {'key': 'pageSize', 'type': 'object'}, - 'preferred_regions': {'key': 'preferredRegions', 'type': 'object'}, - 'detect_datetime': {'key': 'detectDatetime', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CosmosDbSqlApiSource, self).__init__(**kwargs) - self.type = 'CosmosDbSqlApiSource' # type: str - self.query = kwargs.get('query', None) - self.page_size = kwargs.get('page_size', None) - self.preferred_regions = kwargs.get('preferred_regions', None) - self.detect_datetime = kwargs.get('detect_datetime', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class CouchbaseLinkedService(LinkedService): - """Couchbase server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: An ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param cred_string: The Azure key vault secret reference of credString in connection string. - :type cred_string: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'cred_string': {'key': 'typeProperties.credString', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CouchbaseLinkedService, self).__init__(**kwargs) - self.type = 'Couchbase' # type: str - self.connection_string = kwargs.get('connection_string', None) - self.cred_string = kwargs.get('cred_string', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class CouchbaseSource(TabularSource): - """A copy activity Couchbase server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CouchbaseSource, self).__init__(**kwargs) - self.type = 'CouchbaseSource' # type: str - self.query = kwargs.get('query', None) - - -class CouchbaseTableDataset(Dataset): - """Couchbase server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CouchbaseTableDataset, self).__init__(**kwargs) - self.type = 'CouchbaseTable' # type: str - self.table_name = kwargs.get('table_name', None) - - -class CreateDataFlowDebugSessionRequest(msrest.serialization.Model): - """Request body structure for creating data flow debug session. - - :param data_flow_name: The name of the data flow. - :type data_flow_name: str - :param existing_cluster_id: The ID of existing Databricks cluster. - :type existing_cluster_id: str - :param cluster_timeout: Timeout setting for Databricks cluster. - :type cluster_timeout: int - :param new_cluster_name: The name of new Databricks cluster. - :type new_cluster_name: str - :param new_cluster_node_type: The type of new Databricks cluster. - :type new_cluster_node_type: str - :param data_bricks_linked_service: Data bricks linked service. - :type data_bricks_linked_service: - ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceResource - """ - - _attribute_map = { - 'data_flow_name': {'key': 'dataFlowName', 'type': 'str'}, - 'existing_cluster_id': {'key': 'existingClusterId', 'type': 'str'}, - 'cluster_timeout': {'key': 'clusterTimeout', 'type': 'int'}, - 'new_cluster_name': {'key': 'newClusterName', 'type': 'str'}, - 'new_cluster_node_type': {'key': 'newClusterNodeType', 'type': 'str'}, - 'data_bricks_linked_service': {'key': 'dataBricksLinkedService', 'type': 'LinkedServiceResource'}, - } - - def __init__( - self, - **kwargs - ): - super(CreateDataFlowDebugSessionRequest, self).__init__(**kwargs) - self.data_flow_name = kwargs.get('data_flow_name', None) - self.existing_cluster_id = kwargs.get('existing_cluster_id', None) - self.cluster_timeout = kwargs.get('cluster_timeout', None) - self.new_cluster_name = kwargs.get('new_cluster_name', None) - self.new_cluster_node_type = kwargs.get('new_cluster_node_type', None) - self.data_bricks_linked_service = kwargs.get('data_bricks_linked_service', None) - - -class CreateDataFlowDebugSessionResponse(msrest.serialization.Model): - """Response body structure for creating data flow debug session. - - :param session_id: The ID of data flow debug session. - :type session_id: str - """ - - _attribute_map = { - 'session_id': {'key': 'sessionId', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(CreateDataFlowDebugSessionResponse, self).__init__(**kwargs) - self.session_id = kwargs.get('session_id', None) - - -class CreateRunResponse(msrest.serialization.Model): - """Response body with a run identifier. - - All required parameters must be populated in order to send to Azure. - - :param run_id: Required. Identifier of a run. - :type run_id: str - """ - - _validation = { - 'run_id': {'required': True}, - } - - _attribute_map = { - 'run_id': {'key': 'runId', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(CreateRunResponse, self).__init__(**kwargs) - self.run_id = kwargs['run_id'] - - -class CustomActivity(ExecutionActivity): - """Custom activity type. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param command: Required. Command for custom activity Type: string (or Expression with - resultType string). - :type command: any - :param resource_linked_service: Resource linked service reference. - :type resource_linked_service: - ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param folder_path: Folder path for resource files Type: string (or Expression with resultType - string). - :type folder_path: any - :param reference_objects: Reference objects. - :type reference_objects: - ~azure.synapse.artifacts.v2020_12_01.models.CustomActivityReferenceObject - :param extended_properties: User defined property bag. There is no restriction on the keys or - values that can be used. The user specified custom activity has the full responsibility to - consume and interpret the content defined. - :type extended_properties: dict[str, any] - :param retention_time_in_days: The retention time for the files submitted for custom activity. - Type: double (or Expression with resultType double). - :type retention_time_in_days: any - :param auto_user_specification: Elevation level and scope for the user, default is nonadmin - task. Type: string (or Expression with resultType double). - :type auto_user_specification: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'command': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'command': {'key': 'typeProperties.command', 'type': 'object'}, - 'resource_linked_service': {'key': 'typeProperties.resourceLinkedService', 'type': 'LinkedServiceReference'}, - 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, - 'reference_objects': {'key': 'typeProperties.referenceObjects', 'type': 'CustomActivityReferenceObject'}, - 'extended_properties': {'key': 'typeProperties.extendedProperties', 'type': '{object}'}, - 'retention_time_in_days': {'key': 'typeProperties.retentionTimeInDays', 'type': 'object'}, - 'auto_user_specification': {'key': 'typeProperties.autoUserSpecification', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CustomActivity, self).__init__(**kwargs) - self.type = 'Custom' # type: str - self.command = kwargs['command'] - self.resource_linked_service = kwargs.get('resource_linked_service', None) - self.folder_path = kwargs.get('folder_path', None) - self.reference_objects = kwargs.get('reference_objects', None) - self.extended_properties = kwargs.get('extended_properties', None) - self.retention_time_in_days = kwargs.get('retention_time_in_days', None) - self.auto_user_specification = kwargs.get('auto_user_specification', None) - - -class CustomActivityReferenceObject(msrest.serialization.Model): - """Reference objects for custom activity. - - :param linked_services: Linked service references. - :type linked_services: list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference] - :param datasets: Dataset references. - :type datasets: list[~azure.synapse.artifacts.v2020_12_01.models.DatasetReference] - """ - - _attribute_map = { - 'linked_services': {'key': 'linkedServices', 'type': '[LinkedServiceReference]'}, - 'datasets': {'key': 'datasets', 'type': '[DatasetReference]'}, - } - - def __init__( - self, - **kwargs - ): - super(CustomActivityReferenceObject, self).__init__(**kwargs) - self.linked_services = kwargs.get('linked_services', None) - self.datasets = kwargs.get('datasets', None) - - -class CustomDataset(Dataset): - """The custom dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param type_properties: Custom dataset properties. - :type type_properties: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'type_properties': {'key': 'typeProperties', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CustomDataset, self).__init__(**kwargs) - self.type = 'CustomDataset' # type: str - self.type_properties = kwargs.get('type_properties', None) - - -class CustomDataSourceLinkedService(LinkedService): - """Custom linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param type_properties: Required. Custom linked service properties. - :type type_properties: any - """ - - _validation = { - 'type': {'required': True}, - 'type_properties': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'type_properties': {'key': 'typeProperties', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(CustomDataSourceLinkedService, self).__init__(**kwargs) - self.type = 'CustomDataSource' # type: str - self.type_properties = kwargs['type_properties'] - - -class CustomerManagedKeyDetails(msrest.serialization.Model): - """Details of the customer managed key associated with the workspace. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar status: The customer managed key status on the workspace. - :vartype status: str - :param key: The key object of the workspace. - :type key: ~azure.synapse.artifacts.v2020_12_01.models.WorkspaceKeyDetails - """ - - _validation = { - 'status': {'readonly': True}, - } - - _attribute_map = { - 'status': {'key': 'status', 'type': 'str'}, - 'key': {'key': 'key', 'type': 'WorkspaceKeyDetails'}, - } - - def __init__( - self, - **kwargs - ): - super(CustomerManagedKeyDetails, self).__init__(**kwargs) - self.status = None - self.key = kwargs.get('key', None) - - -class CustomEventsTrigger(MultiplePipelineTrigger): - """Trigger that runs every time a custom event is received. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Trigger type.Constant filled by server. - :type type: str - :param description: Trigger description. - :type description: str - :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are - called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRuntimeState - :param annotations: List of tags that can be used for describing the trigger. - :type annotations: list[any] - :param pipelines: Pipelines that need to be started. - :type pipelines: list[~azure.synapse.artifacts.v2020_12_01.models.TriggerPipelineReference] - :param subject_begins_with: The event subject must begin with the pattern provided for trigger - to fire. At least one of these must be provided: subjectBeginsWith, subjectEndsWith. - :type subject_begins_with: str - :param subject_ends_with: The event subject must end with the pattern provided for trigger to - fire. At least one of these must be provided: subjectBeginsWith, subjectEndsWith. - :type subject_ends_with: str - :param events: Required. The list of event types that cause this trigger to fire. - :type events: list[any] - :param scope: Required. The ARM resource ID of the Azure Event Grid Topic. - :type scope: str - """ - - _validation = { - 'type': {'required': True}, - 'runtime_state': {'readonly': True}, - 'events': {'required': True}, - 'scope': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, - 'subject_begins_with': {'key': 'typeProperties.subjectBeginsWith', 'type': 'str'}, - 'subject_ends_with': {'key': 'typeProperties.subjectEndsWith', 'type': 'str'}, - 'events': {'key': 'typeProperties.events', 'type': '[object]'}, - 'scope': {'key': 'typeProperties.scope', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(CustomEventsTrigger, self).__init__(**kwargs) - self.type = 'CustomEventsTrigger' # type: str - self.subject_begins_with = kwargs.get('subject_begins_with', None) - self.subject_ends_with = kwargs.get('subject_ends_with', None) - self.events = kwargs['events'] - self.scope = kwargs['scope'] - - -class CustomSetupBase(msrest.serialization.Model): - """The base definition of the custom setup. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: . - - All required parameters must be populated in order to send to Azure. - - :param type: Required. The type of custom setup.Constant filled by server. - :type type: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - } - - _subtype_map = { - 'type': {} - } - - def __init__( - self, - **kwargs - ): - super(CustomSetupBase, self).__init__(**kwargs) - self.type = None # type: Optional[str] - - -class DatabricksNotebookActivity(ExecutionActivity): - """DatabricksNotebook activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param notebook_path: Required. The absolute path of the notebook to be run in the Databricks - Workspace. This path must begin with a slash. Type: string (or Expression with resultType - string). - :type notebook_path: any - :param base_parameters: Base parameters to be used for each run of this job.If the notebook - takes a parameter that is not specified, the default value from the notebook will be used. - :type base_parameters: dict[str, any] - :param libraries: A list of libraries to be installed on the cluster that will execute the job. - :type libraries: list[dict[str, any]] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'notebook_path': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'notebook_path': {'key': 'typeProperties.notebookPath', 'type': 'object'}, - 'base_parameters': {'key': 'typeProperties.baseParameters', 'type': '{object}'}, - 'libraries': {'key': 'typeProperties.libraries', 'type': '[{object}]'}, - } - - def __init__( - self, - **kwargs - ): - super(DatabricksNotebookActivity, self).__init__(**kwargs) - self.type = 'DatabricksNotebook' # type: str - self.notebook_path = kwargs['notebook_path'] - self.base_parameters = kwargs.get('base_parameters', None) - self.libraries = kwargs.get('libraries', None) - - -class DatabricksSparkJarActivity(ExecutionActivity): - """DatabricksSparkJar activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param main_class_name: Required. The full name of the class containing the main method to be - executed. This class must be contained in a JAR provided as a library. Type: string (or - Expression with resultType string). - :type main_class_name: any - :param parameters: Parameters that will be passed to the main method. - :type parameters: list[any] - :param libraries: A list of libraries to be installed on the cluster that will execute the job. - :type libraries: list[dict[str, any]] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'main_class_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'main_class_name': {'key': 'typeProperties.mainClassName', 'type': 'object'}, - 'parameters': {'key': 'typeProperties.parameters', 'type': '[object]'}, - 'libraries': {'key': 'typeProperties.libraries', 'type': '[{object}]'}, - } - - def __init__( - self, - **kwargs - ): - super(DatabricksSparkJarActivity, self).__init__(**kwargs) - self.type = 'DatabricksSparkJar' # type: str - self.main_class_name = kwargs['main_class_name'] - self.parameters = kwargs.get('parameters', None) - self.libraries = kwargs.get('libraries', None) - - -class DatabricksSparkPythonActivity(ExecutionActivity): - """DatabricksSparkPython activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param python_file: Required. The URI of the Python file to be executed. DBFS paths are - supported. Type: string (or Expression with resultType string). - :type python_file: any - :param parameters: Command line parameters that will be passed to the Python file. - :type parameters: list[any] - :param libraries: A list of libraries to be installed on the cluster that will execute the job. - :type libraries: list[dict[str, any]] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'python_file': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'python_file': {'key': 'typeProperties.pythonFile', 'type': 'object'}, - 'parameters': {'key': 'typeProperties.parameters', 'type': '[object]'}, - 'libraries': {'key': 'typeProperties.libraries', 'type': '[{object}]'}, - } - - def __init__( - self, - **kwargs - ): - super(DatabricksSparkPythonActivity, self).__init__(**kwargs) - self.type = 'DatabricksSparkPython' # type: str - self.python_file = kwargs['python_file'] - self.parameters = kwargs.get('parameters', None) - self.libraries = kwargs.get('libraries', None) - - -class DataFlow(msrest.serialization.Model): - """Azure Synapse nested object which contains a flow with data movements and transformations. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: MappingDataFlow. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Type of data flow.Constant filled by server. - :type type: str - :param description: The description of the data flow. - :type description: str - :param annotations: List of tags that can be used for describing the data flow. - :type annotations: list[any] - :param folder: The folder that this data flow is in. If not specified, Data flow will appear at - the root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowFolder - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DataFlowFolder'}, - } - - _subtype_map = { - 'type': {'MappingDataFlow': 'MappingDataFlow'} - } - - def __init__( - self, - **kwargs - ): - super(DataFlow, self).__init__(**kwargs) - self.type = None # type: Optional[str] - self.description = kwargs.get('description', None) - self.annotations = kwargs.get('annotations', None) - self.folder = kwargs.get('folder', None) - - -class DataFlowDebugCommandRequest(msrest.serialization.Model): - """Request body structure for data flow expression preview. - - All required parameters must be populated in order to send to Azure. - - :param session_id: Required. The ID of data flow debug session. - :type session_id: str - :param data_flow_name: The data flow which contains the debug session. - :type data_flow_name: str - :param command_name: The command name. - :type command_name: str - :param command_payload: Required. The command payload object. - :type command_payload: any - """ - - _validation = { - 'session_id': {'required': True}, - 'command_payload': {'required': True}, - } - - _attribute_map = { - 'session_id': {'key': 'sessionId', 'type': 'str'}, - 'data_flow_name': {'key': 'dataFlowName', 'type': 'str'}, - 'command_name': {'key': 'commandName', 'type': 'str'}, - 'command_payload': {'key': 'commandPayload', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DataFlowDebugCommandRequest, self).__init__(**kwargs) - self.session_id = kwargs['session_id'] - self.data_flow_name = kwargs.get('data_flow_name', None) - self.command_name = kwargs.get('command_name', None) - self.command_payload = kwargs['command_payload'] - - -class DataFlowDebugCommandResponse(msrest.serialization.Model): - """Response body structure of data flow result for data preview, statistics or expression preview. - - :param status: The run status of data preview, statistics or expression preview. - :type status: str - :param data: The result data of data preview, statistics or expression preview. - :type data: str - """ - - _attribute_map = { - 'status': {'key': 'status', 'type': 'str'}, - 'data': {'key': 'data', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(DataFlowDebugCommandResponse, self).__init__(**kwargs) - self.status = kwargs.get('status', None) - self.data = kwargs.get('data', None) - - -class DataFlowDebugPackage(msrest.serialization.Model): - """Request body structure for starting data flow debug session. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param session_id: The ID of data flow debug session. - :type session_id: str - :param data_flow: Data flow instance. - :type data_flow: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowDebugResource - :param datasets: List of datasets. - :type datasets: list[~azure.synapse.artifacts.v2020_12_01.models.DatasetDebugResource] - :param linked_services: List of linked services. - :type linked_services: - list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceDebugResource] - :param staging: Staging info for debug session. - :type staging: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowStagingInfo - :param debug_settings: Data flow debug settings. - :type debug_settings: - ~azure.synapse.artifacts.v2020_12_01.models.DataFlowDebugPackageDebugSettings - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'session_id': {'key': 'sessionId', 'type': 'str'}, - 'data_flow': {'key': 'dataFlow', 'type': 'DataFlowDebugResource'}, - 'datasets': {'key': 'datasets', 'type': '[DatasetDebugResource]'}, - 'linked_services': {'key': 'linkedServices', 'type': '[LinkedServiceDebugResource]'}, - 'staging': {'key': 'staging', 'type': 'DataFlowStagingInfo'}, - 'debug_settings': {'key': 'debugSettings', 'type': 'DataFlowDebugPackageDebugSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(DataFlowDebugPackage, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.session_id = kwargs.get('session_id', None) - self.data_flow = kwargs.get('data_flow', None) - self.datasets = kwargs.get('datasets', None) - self.linked_services = kwargs.get('linked_services', None) - self.staging = kwargs.get('staging', None) - self.debug_settings = kwargs.get('debug_settings', None) - - -class DataFlowDebugPackageDebugSettings(msrest.serialization.Model): - """Data flow debug settings. - - :param source_settings: Source setting for data flow debug. - :type source_settings: list[~azure.synapse.artifacts.v2020_12_01.models.DataFlowSourceSetting] - :param parameters: Data flow parameters. - :type parameters: dict[str, any] - :param dataset_parameters: Parameters for dataset. - :type dataset_parameters: any - """ - - _attribute_map = { - 'source_settings': {'key': 'sourceSettings', 'type': '[DataFlowSourceSetting]'}, - 'parameters': {'key': 'parameters', 'type': '{object}'}, - 'dataset_parameters': {'key': 'datasetParameters', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DataFlowDebugPackageDebugSettings, self).__init__(**kwargs) - self.source_settings = kwargs.get('source_settings', None) - self.parameters = kwargs.get('parameters', None) - self.dataset_parameters = kwargs.get('dataset_parameters', None) - - -class DataFlowDebugPreviewDataRequest(msrest.serialization.Model): - """Request body structure for data flow preview data. - - :param session_id: The ID of data flow debug session. - :type session_id: str - :param data_flow_name: The data flow which contains the debug session. - :type data_flow_name: str - :param stream_name: The output stream name. - :type stream_name: str - :param row_limits: The row limit for preview request. - :type row_limits: int - """ - - _attribute_map = { - 'session_id': {'key': 'sessionId', 'type': 'str'}, - 'data_flow_name': {'key': 'dataFlowName', 'type': 'str'}, - 'stream_name': {'key': 'streamName', 'type': 'str'}, - 'row_limits': {'key': 'rowLimits', 'type': 'int'}, - } - - def __init__( - self, - **kwargs - ): - super(DataFlowDebugPreviewDataRequest, self).__init__(**kwargs) - self.session_id = kwargs.get('session_id', None) - self.data_flow_name = kwargs.get('data_flow_name', None) - self.stream_name = kwargs.get('stream_name', None) - self.row_limits = kwargs.get('row_limits', None) - - -class DataFlowDebugQueryResponse(msrest.serialization.Model): - """Response body structure of data flow query for data preview, statistics or expression preview. - - :param run_id: The run ID of data flow debug session. - :type run_id: str - """ - - _attribute_map = { - 'run_id': {'key': 'runId', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(DataFlowDebugQueryResponse, self).__init__(**kwargs) - self.run_id = kwargs.get('run_id', None) - - -class SubResourceDebugResource(msrest.serialization.Model): - """Azure Synapse nested debug resource. - - :param name: The resource name. - :type name: str - """ - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SubResourceDebugResource, self).__init__(**kwargs) - self.name = kwargs.get('name', None) - - -class DataFlowDebugResource(SubResourceDebugResource): - """Data flow debug resource. - - All required parameters must be populated in order to send to Azure. - - :param name: The resource name. - :type name: str - :param properties: Required. Data flow properties. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.DataFlow - """ - - _validation = { - 'properties': {'required': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'DataFlow'}, - } - - def __init__( - self, - **kwargs - ): - super(DataFlowDebugResource, self).__init__(**kwargs) - self.properties = kwargs['properties'] - - -class DataFlowDebugResultResponse(msrest.serialization.Model): - """Response body structure of data flow result for data preview, statistics or expression preview. - - :param status: The run status of data preview, statistics or expression preview. - :type status: str - :param data: The result data of data preview, statistics or expression preview. - :type data: str - """ - - _attribute_map = { - 'status': {'key': 'status', 'type': 'str'}, - 'data': {'key': 'data', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(DataFlowDebugResultResponse, self).__init__(**kwargs) - self.status = kwargs.get('status', None) - self.data = kwargs.get('data', None) - - -class DataFlowDebugSessionInfo(msrest.serialization.Model): - """Data flow debug session info. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param data_flow_name: The name of the data flow. - :type data_flow_name: str - :param compute_type: Compute type of the cluster. - :type compute_type: str - :param core_count: Core count of the cluster. - :type core_count: int - :param node_count: Node count of the cluster. (deprecated property). - :type node_count: int - :param integration_runtime_name: Attached integration runtime name of data flow debug session. - :type integration_runtime_name: str - :param session_id: The ID of data flow debug session. - :type session_id: str - :param start_time: Start time of data flow debug session. - :type start_time: str - :param time_to_live_in_minutes: Compute type of the cluster. - :type time_to_live_in_minutes: int - :param last_activity_time: Last activity time of data flow debug session. - :type last_activity_time: str - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'data_flow_name': {'key': 'dataFlowName', 'type': 'str'}, - 'compute_type': {'key': 'computeType', 'type': 'str'}, - 'core_count': {'key': 'coreCount', 'type': 'int'}, - 'node_count': {'key': 'nodeCount', 'type': 'int'}, - 'integration_runtime_name': {'key': 'integrationRuntimeName', 'type': 'str'}, - 'session_id': {'key': 'sessionId', 'type': 'str'}, - 'start_time': {'key': 'startTime', 'type': 'str'}, - 'time_to_live_in_minutes': {'key': 'timeToLiveInMinutes', 'type': 'int'}, - 'last_activity_time': {'key': 'lastActivityTime', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(DataFlowDebugSessionInfo, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.data_flow_name = kwargs.get('data_flow_name', None) - self.compute_type = kwargs.get('compute_type', None) - self.core_count = kwargs.get('core_count', None) - self.node_count = kwargs.get('node_count', None) - self.integration_runtime_name = kwargs.get('integration_runtime_name', None) - self.session_id = kwargs.get('session_id', None) - self.start_time = kwargs.get('start_time', None) - self.time_to_live_in_minutes = kwargs.get('time_to_live_in_minutes', None) - self.last_activity_time = kwargs.get('last_activity_time', None) - - -class DataFlowDebugStatisticsRequest(msrest.serialization.Model): - """Request body structure for data flow statistics. - - :param session_id: The ID of data flow debug session. - :type session_id: str - :param data_flow_name: The data flow which contains the debug session. - :type data_flow_name: str - :param stream_name: The output stream name. - :type stream_name: str - :param columns: List of column names. - :type columns: list[str] - """ - - _attribute_map = { - 'session_id': {'key': 'sessionId', 'type': 'str'}, - 'data_flow_name': {'key': 'dataFlowName', 'type': 'str'}, - 'stream_name': {'key': 'streamName', 'type': 'str'}, - 'columns': {'key': 'columns', 'type': '[str]'}, - } - - def __init__( - self, - **kwargs - ): - super(DataFlowDebugStatisticsRequest, self).__init__(**kwargs) - self.session_id = kwargs.get('session_id', None) - self.data_flow_name = kwargs.get('data_flow_name', None) - self.stream_name = kwargs.get('stream_name', None) - self.columns = kwargs.get('columns', None) - - -class DataFlowFolder(msrest.serialization.Model): - """The folder that this data flow is in. If not specified, Data flow will appear at the root level. - - :param name: The name of the folder that this data flow is in. - :type name: str - """ - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(DataFlowFolder, self).__init__(**kwargs) - self.name = kwargs.get('name', None) - - -class DataFlowListResponse(msrest.serialization.Model): - """A list of data flow resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of data flows. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.DataFlowResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[DataFlowResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(DataFlowListResponse, self).__init__(**kwargs) - self.value = kwargs['value'] - self.next_link = kwargs.get('next_link', None) - - -class DataFlowReference(msrest.serialization.Model): - """Data flow reference type. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Data flow reference type. Possible values include: "DataFlowReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.DataFlowReferenceType - :param reference_name: Required. Reference data flow name. - :type reference_name: str - :param dataset_parameters: Reference data flow parameters from dataset. - :type dataset_parameters: any - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - 'dataset_parameters': {'key': 'datasetParameters', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DataFlowReference, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = kwargs['type'] - self.reference_name = kwargs['reference_name'] - self.dataset_parameters = kwargs.get('dataset_parameters', None) - - -class SubResource(AzureEntityResource): - """Azure Synapse nested resource, which belongs to a workspace. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SubResource, self).__init__(**kwargs) - - -class DataFlowResource(SubResource): - """Data flow resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Data flow properties. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.DataFlow - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'DataFlow'}, - } - - def __init__( - self, - **kwargs - ): - super(DataFlowResource, self).__init__(**kwargs) - self.properties = kwargs['properties'] - - -class Transformation(msrest.serialization.Model): - """A data flow transformation. - - All required parameters must be populated in order to send to Azure. - - :param name: Required. Transformation name. - :type name: str - :param description: Transformation description. - :type description: str - """ - - _validation = { - 'name': {'required': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(Transformation, self).__init__(**kwargs) - self.name = kwargs['name'] - self.description = kwargs.get('description', None) - - -class DataFlowSink(Transformation): - """Transformation for data flow sink. - - All required parameters must be populated in order to send to Azure. - - :param name: Required. Transformation name. - :type name: str - :param description: Transformation description. - :type description: str - :param dataset: Dataset reference. - :type dataset: ~azure.synapse.artifacts.v2020_12_01.models.DatasetReference - :param linked_service: Linked service reference. - :type linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param schema_linked_service: Schema linked service reference. - :type schema_linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - """ - - _validation = { - 'name': {'required': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'dataset': {'key': 'dataset', 'type': 'DatasetReference'}, - 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, - 'schema_linked_service': {'key': 'schemaLinkedService', 'type': 'LinkedServiceReference'}, - } - - def __init__( - self, - **kwargs - ): - super(DataFlowSink, self).__init__(**kwargs) - self.dataset = kwargs.get('dataset', None) - self.linked_service = kwargs.get('linked_service', None) - self.schema_linked_service = kwargs.get('schema_linked_service', None) - - -class DataFlowSource(Transformation): - """Transformation for data flow source. - - All required parameters must be populated in order to send to Azure. - - :param name: Required. Transformation name. - :type name: str - :param description: Transformation description. - :type description: str - :param dataset: Dataset reference. - :type dataset: ~azure.synapse.artifacts.v2020_12_01.models.DatasetReference - :param linked_service: Linked service reference. - :type linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param schema_linked_service: Schema linked service reference. - :type schema_linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - """ - - _validation = { - 'name': {'required': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'dataset': {'key': 'dataset', 'type': 'DatasetReference'}, - 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, - 'schema_linked_service': {'key': 'schemaLinkedService', 'type': 'LinkedServiceReference'}, - } - - def __init__( - self, - **kwargs - ): - super(DataFlowSource, self).__init__(**kwargs) - self.dataset = kwargs.get('dataset', None) - self.linked_service = kwargs.get('linked_service', None) - self.schema_linked_service = kwargs.get('schema_linked_service', None) - - -class DataFlowSourceSetting(msrest.serialization.Model): - """Definition of data flow source setting for debug. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param source_name: The data flow source name. - :type source_name: str - :param row_limit: Defines the row limit of data flow source in debug. - :type row_limit: int - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'source_name': {'key': 'sourceName', 'type': 'str'}, - 'row_limit': {'key': 'rowLimit', 'type': 'int'}, - } - - def __init__( - self, - **kwargs - ): - super(DataFlowSourceSetting, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.source_name = kwargs.get('source_name', None) - self.row_limit = kwargs.get('row_limit', None) - - -class DataFlowStagingInfo(msrest.serialization.Model): - """Staging info for execute data flow activity. - - :param linked_service: Staging linked service reference. - :type linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param folder_path: Folder path for staging blob. - :type folder_path: str - """ - - _attribute_map = { - 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, - 'folder_path': {'key': 'folderPath', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(DataFlowStagingInfo, self).__init__(**kwargs) - self.linked_service = kwargs.get('linked_service', None) - self.folder_path = kwargs.get('folder_path', None) - - -class DataLakeAnalyticsUSQLActivity(ExecutionActivity): - """Data Lake Analytics U-SQL activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param script_path: Required. Case-sensitive path to folder that contains the U-SQL script. - Type: string (or Expression with resultType string). - :type script_path: any - :param script_linked_service: Required. Script linked service reference. - :type script_linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param degree_of_parallelism: The maximum number of nodes simultaneously used to run the job. - Default value is 1. Type: integer (or Expression with resultType integer), minimum: 1. - :type degree_of_parallelism: any - :param priority: Determines which jobs out of all that are queued should be selected to run - first. The lower the number, the higher the priority. Default value is 1000. Type: integer (or - Expression with resultType integer), minimum: 1. - :type priority: any - :param parameters: Parameters for U-SQL job request. - :type parameters: dict[str, any] - :param runtime_version: Runtime version of the U-SQL engine to use. Type: string (or Expression - with resultType string). - :type runtime_version: any - :param compilation_mode: Compilation mode of U-SQL. Must be one of these values : Semantic, - Full and SingleBox. Type: string (or Expression with resultType string). - :type compilation_mode: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'script_path': {'required': True}, - 'script_linked_service': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'script_path': {'key': 'typeProperties.scriptPath', 'type': 'object'}, - 'script_linked_service': {'key': 'typeProperties.scriptLinkedService', 'type': 'LinkedServiceReference'}, - 'degree_of_parallelism': {'key': 'typeProperties.degreeOfParallelism', 'type': 'object'}, - 'priority': {'key': 'typeProperties.priority', 'type': 'object'}, - 'parameters': {'key': 'typeProperties.parameters', 'type': '{object}'}, - 'runtime_version': {'key': 'typeProperties.runtimeVersion', 'type': 'object'}, - 'compilation_mode': {'key': 'typeProperties.compilationMode', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DataLakeAnalyticsUSQLActivity, self).__init__(**kwargs) - self.type = 'DataLakeAnalyticsU-SQL' # type: str - self.script_path = kwargs['script_path'] - self.script_linked_service = kwargs['script_linked_service'] - self.degree_of_parallelism = kwargs.get('degree_of_parallelism', None) - self.priority = kwargs.get('priority', None) - self.parameters = kwargs.get('parameters', None) - self.runtime_version = kwargs.get('runtime_version', None) - self.compilation_mode = kwargs.get('compilation_mode', None) - - -class DataLakeStorageAccountDetails(msrest.serialization.Model): - """Details of the data lake storage account associated with the workspace. - - :param account_url: Account URL. - :type account_url: str - :param filesystem: Filesystem name. - :type filesystem: str - """ - - _attribute_map = { - 'account_url': {'key': 'accountUrl', 'type': 'str'}, - 'filesystem': {'key': 'filesystem', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(DataLakeStorageAccountDetails, self).__init__(**kwargs) - self.account_url = kwargs.get('account_url', None) - self.filesystem = kwargs.get('filesystem', None) - - -class DatasetCompression(msrest.serialization.Model): - """The compression method used on a dataset. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: DatasetBZip2Compression, DatasetDeflateCompression, DatasetGZipCompression, DatasetTarCompression, DatasetTarGZipCompression, DatasetZipDeflateCompression. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, - } - - _subtype_map = { - 'type': {'BZip2': 'DatasetBZip2Compression', 'Deflate': 'DatasetDeflateCompression', 'GZip': 'DatasetGZipCompression', 'Tar': 'DatasetTarCompression', 'TarGZip': 'DatasetTarGZipCompression', 'ZipDeflate': 'DatasetZipDeflateCompression'} - } - - def __init__( - self, - **kwargs - ): - super(DatasetCompression, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = 'DatasetCompression' # type: str - - -class DatasetBZip2Compression(DatasetCompression): - """The BZip2 compression method used on a dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DatasetBZip2Compression, self).__init__(**kwargs) - self.type = 'BZip2' # type: str - - -class DatasetDataElement(msrest.serialization.Model): - """Columns that define the structure of the dataset. - - :param name: Name of the column. Type: string (or Expression with resultType string). - :type name: any - :param type: Type of the column. Type: string (or Expression with resultType string). - :type type: any - """ - - _attribute_map = { - 'name': {'key': 'name', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DatasetDataElement, self).__init__(**kwargs) - self.name = kwargs.get('name', None) - self.type = kwargs.get('type', None) - - -class DatasetDebugResource(SubResourceDebugResource): - """Dataset debug resource. - - All required parameters must be populated in order to send to Azure. - - :param name: The resource name. - :type name: str - :param properties: Required. Dataset properties. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.Dataset - """ - - _validation = { - 'properties': {'required': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'Dataset'}, - } - - def __init__( - self, - **kwargs - ): - super(DatasetDebugResource, self).__init__(**kwargs) - self.properties = kwargs['properties'] - - -class DatasetDeflateCompression(DatasetCompression): - """The Deflate compression method used on a dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - :param level: The Deflate compression level. - :type level: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, - 'level': {'key': 'level', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DatasetDeflateCompression, self).__init__(**kwargs) - self.type = 'Deflate' # type: str - self.level = kwargs.get('level', None) - - -class DatasetFolder(msrest.serialization.Model): - """The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - - :param name: The name of the folder that this Dataset is in. - :type name: str - """ - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(DatasetFolder, self).__init__(**kwargs) - self.name = kwargs.get('name', None) - - -class DatasetGZipCompression(DatasetCompression): - """The GZip compression method used on a dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - :param level: The GZip compression level. - :type level: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, - 'level': {'key': 'level', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DatasetGZipCompression, self).__init__(**kwargs) - self.type = 'GZip' # type: str - self.level = kwargs.get('level', None) - - -class DatasetListResponse(msrest.serialization.Model): - """A list of dataset resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of datasets. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.DatasetResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[DatasetResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(DatasetListResponse, self).__init__(**kwargs) - self.value = kwargs['value'] - self.next_link = kwargs.get('next_link', None) - - -class DatasetReference(msrest.serialization.Model): - """Dataset reference type. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Dataset reference type. Possible values include: "DatasetReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.DatasetReferenceType - :param reference_name: Required. Reference dataset name. - :type reference_name: str - :param parameters: Arguments for dataset. - :type parameters: dict[str, any] - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{object}'}, - } - - def __init__( - self, - **kwargs - ): - super(DatasetReference, self).__init__(**kwargs) - self.type = kwargs['type'] - self.reference_name = kwargs['reference_name'] - self.parameters = kwargs.get('parameters', None) - - -class DatasetResource(SubResource): - """Dataset resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Dataset properties. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.Dataset - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'Dataset'}, - } - - def __init__( - self, - **kwargs - ): - super(DatasetResource, self).__init__(**kwargs) - self.properties = kwargs['properties'] - - -class DatasetSchemaDataElement(msrest.serialization.Model): - """Columns that define the physical type schema of the dataset. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Name of the schema column. Type: string (or Expression with resultType string). - :type name: any - :param type: Type of the schema column. Type: string (or Expression with resultType string). - :type type: any - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DatasetSchemaDataElement, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.name = kwargs.get('name', None) - self.type = kwargs.get('type', None) - - -class DatasetTarCompression(DatasetCompression): - """The Tar archive method used on a dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DatasetTarCompression, self).__init__(**kwargs) - self.type = 'Tar' # type: str - - -class DatasetTarGZipCompression(DatasetCompression): - """The TarGZip compression method used on a dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - :param level: The TarGZip compression level. - :type level: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, - 'level': {'key': 'level', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DatasetTarGZipCompression, self).__init__(**kwargs) - self.type = 'TarGZip' # type: str - self.level = kwargs.get('level', None) - - -class DatasetZipDeflateCompression(DatasetCompression): - """The ZipDeflate compression method used on a dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - :param level: The ZipDeflate compression level. - :type level: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, - 'level': {'key': 'level', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DatasetZipDeflateCompression, self).__init__(**kwargs) - self.type = 'ZipDeflate' # type: str - self.level = kwargs.get('level', None) - - -class Db2LinkedService(LinkedService): - """Linked service for DB2 data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: The connection string. It is mutually exclusive with server, - database, authenticationType, userName, packageCollection and certificateCommonName property. - Type: string, SecureString or AzureKeyVaultSecretReference. - :type connection_string: any - :param server: Required. Server name for connection. It is mutually exclusive with - connectionString property. Type: string (or Expression with resultType string). - :type server: any - :param database: Required. Database name for connection. It is mutually exclusive with - connectionString property. Type: string (or Expression with resultType string). - :type database: any - :param authentication_type: AuthenticationType to be used for connection. It is mutually - exclusive with connectionString property. Possible values include: "Basic". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.Db2AuthenticationType - :param username: Username for authentication. It is mutually exclusive with connectionString - property. Type: string (or Expression with resultType string). - :type username: any - :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param package_collection: Under where packages are created when querying database. It is - mutually exclusive with connectionString property. Type: string (or Expression with resultType - string). - :type package_collection: any - :param certificate_common_name: Certificate Common Name when TLS is enabled. It is mutually - exclusive with connectionString property. Type: string (or Expression with resultType string). - :type certificate_common_name: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. It is mutually exclusive with - connectionString property. Type: string (or Expression with resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'server': {'required': True}, - 'database': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'package_collection': {'key': 'typeProperties.packageCollection', 'type': 'object'}, - 'certificate_common_name': {'key': 'typeProperties.certificateCommonName', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(Db2LinkedService, self).__init__(**kwargs) - self.type = 'Db2' # type: str - self.connection_string = kwargs.get('connection_string', None) - self.server = kwargs['server'] - self.database = kwargs['database'] - self.authentication_type = kwargs.get('authentication_type', None) - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.package_collection = kwargs.get('package_collection', None) - self.certificate_common_name = kwargs.get('certificate_common_name', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class Db2Source(TabularSource): - """A copy activity source for Db2 databases. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(Db2Source, self).__init__(**kwargs) - self.type = 'Db2Source' # type: str - self.query = kwargs.get('query', None) - - -class Db2TableDataset(Dataset): - """The Db2 table dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param schema_type_properties_schema: The Db2 schema name. Type: string (or Expression with - resultType string). - :type schema_type_properties_schema: any - :param table: The Db2 table name. Type: string (or Expression with resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(Db2TableDataset, self).__init__(**kwargs) - self.type = 'Db2Table' # type: str - self.table_name = kwargs.get('table_name', None) - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - self.table = kwargs.get('table', None) - - -class DeleteActivity(ExecutionActivity): - """Delete activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param recursive: If true, files or sub-folders under current folder path will be deleted - recursively. Default is false. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param max_concurrent_connections: The max concurrent connections to connect data source at the - same time. - :type max_concurrent_connections: int - :param enable_logging: Whether to record detailed logs of delete-activity execution. Default - value is false. Type: boolean (or Expression with resultType boolean). - :type enable_logging: any - :param log_storage_settings: Log storage settings customer need to provide when enableLogging - is true. - :type log_storage_settings: ~azure.synapse.artifacts.v2020_12_01.models.LogStorageSettings - :param dataset: Required. Delete activity dataset reference. - :type dataset: ~azure.synapse.artifacts.v2020_12_01.models.DatasetReference - :param store_settings: Delete activity store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'max_concurrent_connections': {'minimum': 1}, - 'dataset': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'recursive': {'key': 'typeProperties.recursive', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'typeProperties.maxConcurrentConnections', 'type': 'int'}, - 'enable_logging': {'key': 'typeProperties.enableLogging', 'type': 'object'}, - 'log_storage_settings': {'key': 'typeProperties.logStorageSettings', 'type': 'LogStorageSettings'}, - 'dataset': {'key': 'typeProperties.dataset', 'type': 'DatasetReference'}, - 'store_settings': {'key': 'typeProperties.storeSettings', 'type': 'StoreReadSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(DeleteActivity, self).__init__(**kwargs) - self.type = 'Delete' # type: str - self.recursive = kwargs.get('recursive', None) - self.max_concurrent_connections = kwargs.get('max_concurrent_connections', None) - self.enable_logging = kwargs.get('enable_logging', None) - self.log_storage_settings = kwargs.get('log_storage_settings', None) - self.dataset = kwargs['dataset'] - self.store_settings = kwargs.get('store_settings', None) - - -class DeleteDataFlowDebugSessionRequest(msrest.serialization.Model): - """Request body structure for deleting data flow debug session. - - :param session_id: The ID of data flow debug session. - :type session_id: str - :param data_flow_name: The data flow which contains the debug session. - :type data_flow_name: str - """ - - _attribute_map = { - 'session_id': {'key': 'sessionId', 'type': 'str'}, - 'data_flow_name': {'key': 'dataFlowName', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(DeleteDataFlowDebugSessionRequest, self).__init__(**kwargs) - self.session_id = kwargs.get('session_id', None) - self.data_flow_name = kwargs.get('data_flow_name', None) - - -class DelimitedTextDataset(Dataset): - """Delimited text dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param location: The location of the delimited text storage. - :type location: ~azure.synapse.artifacts.v2020_12_01.models.DatasetLocation - :param column_delimiter: The column delimiter. Type: string (or Expression with resultType - string). - :type column_delimiter: any - :param row_delimiter: The row delimiter. Type: string (or Expression with resultType string). - :type row_delimiter: any - :param encoding_name: The code page name of the preferred encoding. If miss, the default value - is UTF-8, unless BOM denotes another Unicode encoding. Refer to the name column of the table in - the following link to set supported values: - https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string (or Expression with - resultType string). - :type encoding_name: any - :param compression_codec: Possible values include: "bzip2", "gzip", "deflate", "zipDeflate", - "snappy", "lz4", "tar", "tarGZip". - :type compression_codec: str or ~azure.synapse.artifacts.v2020_12_01.models.CompressionCodec - :param compression_level: The data compression method used for DelimitedText. - :type compression_level: any - :param quote_char: The quote character. Type: string (or Expression with resultType string). - :type quote_char: any - :param escape_char: The escape character. Type: string (or Expression with resultType string). - :type escape_char: any - :param first_row_as_header: When used as input, treat the first row of data as headers. When - used as output,write the headers into the output as the first row of data. The default value is - false. Type: boolean (or Expression with resultType boolean). - :type first_row_as_header: any - :param null_value: The null value string. Type: string (or Expression with resultType string). - :type null_value: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'location': {'key': 'typeProperties.location', 'type': 'DatasetLocation'}, - 'column_delimiter': {'key': 'typeProperties.columnDelimiter', 'type': 'object'}, - 'row_delimiter': {'key': 'typeProperties.rowDelimiter', 'type': 'object'}, - 'encoding_name': {'key': 'typeProperties.encodingName', 'type': 'object'}, - 'compression_codec': {'key': 'typeProperties.compressionCodec', 'type': 'str'}, - 'compression_level': {'key': 'typeProperties.compressionLevel', 'type': 'object'}, - 'quote_char': {'key': 'typeProperties.quoteChar', 'type': 'object'}, - 'escape_char': {'key': 'typeProperties.escapeChar', 'type': 'object'}, - 'first_row_as_header': {'key': 'typeProperties.firstRowAsHeader', 'type': 'object'}, - 'null_value': {'key': 'typeProperties.nullValue', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DelimitedTextDataset, self).__init__(**kwargs) - self.type = 'DelimitedText' # type: str - self.location = kwargs.get('location', None) - self.column_delimiter = kwargs.get('column_delimiter', None) - self.row_delimiter = kwargs.get('row_delimiter', None) - self.encoding_name = kwargs.get('encoding_name', None) - self.compression_codec = kwargs.get('compression_codec', None) - self.compression_level = kwargs.get('compression_level', None) - self.quote_char = kwargs.get('quote_char', None) - self.escape_char = kwargs.get('escape_char', None) - self.first_row_as_header = kwargs.get('first_row_as_header', None) - self.null_value = kwargs.get('null_value', None) - - -class DelimitedTextReadSettings(FormatReadSettings): - """Delimited text read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param skip_line_count: Indicates the number of non-empty rows to skip when reading data from - input files. Type: integer (or Expression with resultType integer). - :type skip_line_count: any - :param compression_properties: Compression settings. - :type compression_properties: - ~azure.synapse.artifacts.v2020_12_01.models.CompressionReadSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'skip_line_count': {'key': 'skipLineCount', 'type': 'object'}, - 'compression_properties': {'key': 'compressionProperties', 'type': 'CompressionReadSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(DelimitedTextReadSettings, self).__init__(**kwargs) - self.type = 'DelimitedTextReadSettings' # type: str - self.skip_line_count = kwargs.get('skip_line_count', None) - self.compression_properties = kwargs.get('compression_properties', None) - - -class DelimitedTextSink(CopySink): - """A copy activity DelimitedText sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: DelimitedText store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreWriteSettings - :param format_settings: DelimitedText format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.DelimitedTextWriteSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreWriteSettings'}, - 'format_settings': {'key': 'formatSettings', 'type': 'DelimitedTextWriteSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(DelimitedTextSink, self).__init__(**kwargs) - self.type = 'DelimitedTextSink' # type: str - self.store_settings = kwargs.get('store_settings', None) - self.format_settings = kwargs.get('format_settings', None) - - -class DelimitedTextSource(CopySource): - """A copy activity DelimitedText source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: DelimitedText store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - :param format_settings: DelimitedText format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.DelimitedTextReadSettings - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'format_settings': {'key': 'formatSettings', 'type': 'DelimitedTextReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DelimitedTextSource, self).__init__(**kwargs) - self.type = 'DelimitedTextSource' # type: str - self.store_settings = kwargs.get('store_settings', None) - self.format_settings = kwargs.get('format_settings', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class DelimitedTextWriteSettings(FormatWriteSettings): - """Delimited text write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param quote_all_text: Indicates whether string values should always be enclosed with quotes. - Type: boolean (or Expression with resultType boolean). - :type quote_all_text: any - :param file_extension: Required. The file extension used to create the files. Type: string (or - Expression with resultType string). - :type file_extension: any - :param max_rows_per_file: Limit the written file's row count to be smaller than or equal to the - specified count. Type: integer (or Expression with resultType integer). - :type max_rows_per_file: any - :param file_name_prefix: Specifies the file name pattern - :code:``_:code:``.:code:`` when copy from non-file - based store without partitionOptions. Type: string (or Expression with resultType string). - :type file_name_prefix: any - """ - - _validation = { - 'type': {'required': True}, - 'file_extension': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'quote_all_text': {'key': 'quoteAllText', 'type': 'object'}, - 'file_extension': {'key': 'fileExtension', 'type': 'object'}, - 'max_rows_per_file': {'key': 'maxRowsPerFile', 'type': 'object'}, - 'file_name_prefix': {'key': 'fileNamePrefix', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DelimitedTextWriteSettings, self).__init__(**kwargs) - self.type = 'DelimitedTextWriteSettings' # type: str - self.quote_all_text = kwargs.get('quote_all_text', None) - self.file_extension = kwargs['file_extension'] - self.max_rows_per_file = kwargs.get('max_rows_per_file', None) - self.file_name_prefix = kwargs.get('file_name_prefix', None) - - -class DependencyReference(msrest.serialization.Model): - """Referenced dependency. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: SelfDependencyTumblingWindowTriggerReference, TriggerDependencyReference. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. The type of dependency reference.Constant filled by server. - :type type: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - } - - _subtype_map = { - 'type': {'SelfDependencyTumblingWindowTriggerReference': 'SelfDependencyTumblingWindowTriggerReference', 'TriggerDependencyReference': 'TriggerDependencyReference'} - } - - def __init__( - self, - **kwargs - ): - super(DependencyReference, self).__init__(**kwargs) - self.type = None # type: Optional[str] - - -class DistcpSettings(msrest.serialization.Model): - """Distcp settings. - - All required parameters must be populated in order to send to Azure. - - :param resource_manager_endpoint: Required. Specifies the Yarn ResourceManager endpoint. Type: - string (or Expression with resultType string). - :type resource_manager_endpoint: any - :param temp_script_path: Required. Specifies an existing folder path which will be used to - store temp Distcp command script. The script file is generated by ADF and will be removed after - Copy job finished. Type: string (or Expression with resultType string). - :type temp_script_path: any - :param distcp_options: Specifies the Distcp options. Type: string (or Expression with - resultType string). - :type distcp_options: any - """ - - _validation = { - 'resource_manager_endpoint': {'required': True}, - 'temp_script_path': {'required': True}, - } - - _attribute_map = { - 'resource_manager_endpoint': {'key': 'resourceManagerEndpoint', 'type': 'object'}, - 'temp_script_path': {'key': 'tempScriptPath', 'type': 'object'}, - 'distcp_options': {'key': 'distcpOptions', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DistcpSettings, self).__init__(**kwargs) - self.resource_manager_endpoint = kwargs['resource_manager_endpoint'] - self.temp_script_path = kwargs['temp_script_path'] - self.distcp_options = kwargs.get('distcp_options', None) - - -class DocumentDbCollectionDataset(Dataset): - """Microsoft Azure Document Database Collection dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param collection_name: Required. Document Database collection name. Type: string (or - Expression with resultType string). - :type collection_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'collection_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'collection_name': {'key': 'typeProperties.collectionName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DocumentDbCollectionDataset, self).__init__(**kwargs) - self.type = 'DocumentDbCollection' # type: str - self.collection_name = kwargs['collection_name'] - - -class DocumentDbCollectionSink(CopySink): - """A copy activity Document Database Collection sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param nesting_separator: Nested properties separator. Default is . (dot). Type: string (or - Expression with resultType string). - :type nesting_separator: any - :param write_behavior: Describes how to write data to Azure Cosmos DB. Type: string (or - Expression with resultType string). Allowed values: insert and upsert. - :type write_behavior: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'nesting_separator': {'key': 'nestingSeparator', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DocumentDbCollectionSink, self).__init__(**kwargs) - self.type = 'DocumentDbCollectionSink' # type: str - self.nesting_separator = kwargs.get('nesting_separator', None) - self.write_behavior = kwargs.get('write_behavior', None) - - -class DocumentDbCollectionSource(CopySource): - """A copy activity Document Database Collection source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: Documents query. Type: string (or Expression with resultType string). - :type query: any - :param nesting_separator: Nested properties separator. Type: string (or Expression with - resultType string). - :type nesting_separator: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'nesting_separator': {'key': 'nestingSeparator', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DocumentDbCollectionSource, self).__init__(**kwargs) - self.type = 'DocumentDbCollectionSource' # type: str - self.query = kwargs.get('query', None) - self.nesting_separator = kwargs.get('nesting_separator', None) - self.query_timeout = kwargs.get('query_timeout', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class DrillLinkedService(LinkedService): - """Drill server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: An ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DrillLinkedService, self).__init__(**kwargs) - self.type = 'Drill' # type: str - self.connection_string = kwargs.get('connection_string', None) - self.pwd = kwargs.get('pwd', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class DrillSource(TabularSource): - """A copy activity Drill server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DrillSource, self).__init__(**kwargs) - self.type = 'DrillSource' # type: str - self.query = kwargs.get('query', None) - - -class DrillTableDataset(Dataset): - """Drill server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The table name of the Drill. Type: string (or Expression with resultType string). - :type table: any - :param schema_type_properties_schema: The schema name of the Drill. Type: string (or Expression - with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DrillTableDataset, self).__init__(**kwargs) - self.type = 'DrillTable' # type: str - self.table_name = kwargs.get('table_name', None) - self.table = kwargs.get('table', None) - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - - -class DWCopyCommandDefaultValue(msrest.serialization.Model): - """Default value. - - :param column_name: Column name. Type: object (or Expression with resultType string). - :type column_name: any - :param default_value: The default value of the column. Type: object (or Expression with - resultType string). - :type default_value: any - """ - - _attribute_map = { - 'column_name': {'key': 'columnName', 'type': 'object'}, - 'default_value': {'key': 'defaultValue', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DWCopyCommandDefaultValue, self).__init__(**kwargs) - self.column_name = kwargs.get('column_name', None) - self.default_value = kwargs.get('default_value', None) - - -class DWCopyCommandSettings(msrest.serialization.Model): - """DW Copy Command settings. - - :param default_values: Specifies the default values for each target column in SQL DW. The - default values in the property overwrite the DEFAULT constraint set in the DB, and identity - column cannot have a default value. Type: array of objects (or Expression with resultType array - of objects). - :type default_values: - list[~azure.synapse.artifacts.v2020_12_01.models.DWCopyCommandDefaultValue] - :param additional_options: Additional options directly passed to SQL DW in Copy Command. Type: - key value pairs (value should be string type) (or Expression with resultType object). Example: - "additionalOptions": { "MAXERRORS": "1000", "DATEFORMAT": "'ymd'" }. - :type additional_options: dict[str, str] - """ - - _attribute_map = { - 'default_values': {'key': 'defaultValues', 'type': '[DWCopyCommandDefaultValue]'}, - 'additional_options': {'key': 'additionalOptions', 'type': '{str}'}, - } - - def __init__( - self, - **kwargs - ): - super(DWCopyCommandSettings, self).__init__(**kwargs) - self.default_values = kwargs.get('default_values', None) - self.additional_options = kwargs.get('additional_options', None) - - -class DynamicExecutorAllocation(msrest.serialization.Model): - """Dynamic Executor Allocation Properties. - - :param enabled: Indicates whether Dynamic Executor Allocation is enabled or not. - :type enabled: bool - """ - - _attribute_map = { - 'enabled': {'key': 'enabled', 'type': 'bool'}, - } - - def __init__( - self, - **kwargs - ): - super(DynamicExecutorAllocation, self).__init__(**kwargs) - self.enabled = kwargs.get('enabled', None) - - -class DynamicsAXLinkedService(LinkedService): - """Dynamics AX linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param url: Required. The Dynamics AX (or Dynamics 365 Finance and Operations) instance OData - endpoint. - :type url: any - :param service_principal_id: Required. Specify the application's client ID. Type: string (or - Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: Required. Specify the application's key. Mark this field as a - SecureString to store it securely in Data Factory, or reference a secret stored in Azure Key - Vault. Type: string (or Expression with resultType string). - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: Required. Specify the tenant information (domain name or tenant ID) under which - your application resides. Retrieve it by hovering the mouse in the top-right corner of the - Azure portal. Type: string (or Expression with resultType string). - :type tenant: any - :param aad_resource_id: Required. Specify the resource you are requesting authorization. Type: - string (or Expression with resultType string). - :type aad_resource_id: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - 'service_principal_id': {'required': True}, - 'service_principal_key': {'required': True}, - 'tenant': {'required': True}, - 'aad_resource_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'aad_resource_id': {'key': 'typeProperties.aadResourceId', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DynamicsAXLinkedService, self).__init__(**kwargs) - self.type = 'DynamicsAX' # type: str - self.url = kwargs['url'] - self.service_principal_id = kwargs['service_principal_id'] - self.service_principal_key = kwargs['service_principal_key'] - self.tenant = kwargs['tenant'] - self.aad_resource_id = kwargs['aad_resource_id'] - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class DynamicsAXResourceDataset(Dataset): - """The path of the Dynamics AX OData entity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param path: Required. The path of the Dynamics AX OData entity. Type: string (or Expression - with resultType string). - :type path: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'path': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'path': {'key': 'typeProperties.path', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DynamicsAXResourceDataset, self).__init__(**kwargs) - self.type = 'DynamicsAXResource' # type: str - self.path = kwargs['path'] - - -class DynamicsAXSource(TabularSource): - """A copy activity Dynamics AX source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - :param http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout - to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DynamicsAXSource, self).__init__(**kwargs) - self.type = 'DynamicsAXSource' # type: str - self.query = kwargs.get('query', None) - self.http_request_timeout = kwargs.get('http_request_timeout', None) - - -class DynamicsCrmEntityDataset(Dataset): - """The Dynamics CRM entity dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param entity_name: The logical name of the entity. Type: string (or Expression with resultType - string). - :type entity_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'entity_name': {'key': 'typeProperties.entityName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DynamicsCrmEntityDataset, self).__init__(**kwargs) - self.type = 'DynamicsCrmEntity' # type: str - self.entity_name = kwargs.get('entity_name', None) - - -class DynamicsCrmLinkedService(LinkedService): - """Dynamics CRM linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param deployment_type: Required. The deployment type of the Dynamics CRM instance. 'Online' - for Dynamics CRM Online and 'OnPremisesWithIfd' for Dynamics CRM on-premises with Ifd. Type: - string (or Expression with resultType string). Possible values include: "Online", - "OnPremisesWithIfd". - :type deployment_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsDeploymentType - :param host_name: The host name of the on-premises Dynamics CRM server. The property is - required for on-prem and not allowed for online. Type: string (or Expression with resultType - string). - :type host_name: any - :param port: The port of on-premises Dynamics CRM server. The property is required for on-prem - and not allowed for online. Default is 443. Type: integer (or Expression with resultType - integer), minimum: 0. - :type port: any - :param service_uri: The URL to the Microsoft Dynamics CRM server. The property is required for - on-line and not allowed for on-prem. Type: string (or Expression with resultType string). - :type service_uri: any - :param organization_name: The organization name of the Dynamics CRM instance. The property is - required for on-prem and required for online when there are more than one Dynamics CRM - instances associated with the user. Type: string (or Expression with resultType string). - :type organization_name: any - :param authentication_type: Required. The authentication type to connect to Dynamics CRM - server. 'Office365' for online scenario, 'Ifd' for on-premises with Ifd scenario, - 'AADServicePrincipal' for Server-To-Server authentication in online scenario. Type: string (or - Expression with resultType string). Possible values include: "Office365", "Ifd", - "AADServicePrincipal". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsAuthenticationType - :param username: User name to access the Dynamics CRM instance. Type: string (or Expression - with resultType string). - :type username: any - :param password: Password to access the Dynamics CRM instance. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param service_principal_id: The client ID of the application in Azure Active Directory used - for Server-To-Server authentication. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_credential_type: A string from ServicePrincipalCredentialEnum or an - expression. - :type service_principal_credential_type: any - :param service_principal_credential: The credential of the service principal object in Azure - Active Directory. If servicePrincipalCredentialType is 'ServicePrincipalKey', - servicePrincipalCredential can be SecureString or AzureKeyVaultSecretReference. If - servicePrincipalCredentialType is 'ServicePrincipalCert', servicePrincipalCredential can only - be AzureKeyVaultSecretReference. - :type service_principal_credential: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'deployment_type': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'str'}, - 'host_name': {'key': 'typeProperties.hostName', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'service_uri': {'key': 'typeProperties.serviceUri', 'type': 'object'}, - 'organization_name': {'key': 'typeProperties.organizationName', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_credential_type': {'key': 'typeProperties.servicePrincipalCredentialType', 'type': 'object'}, - 'service_principal_credential': {'key': 'typeProperties.servicePrincipalCredential', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DynamicsCrmLinkedService, self).__init__(**kwargs) - self.type = 'DynamicsCrm' # type: str - self.deployment_type = kwargs['deployment_type'] - self.host_name = kwargs.get('host_name', None) - self.port = kwargs.get('port', None) - self.service_uri = kwargs.get('service_uri', None) - self.organization_name = kwargs.get('organization_name', None) - self.authentication_type = kwargs['authentication_type'] - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.service_principal_id = kwargs.get('service_principal_id', None) - self.service_principal_credential_type = kwargs.get('service_principal_credential_type', None) - self.service_principal_credential = kwargs.get('service_principal_credential', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class DynamicsCrmSink(CopySink): - """A copy activity Dynamics CRM sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param write_behavior: Required. The write behavior for the operation. Possible values include: - "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsSinkWriteBehavior - :param ignore_null_values: The flag indicating whether to ignore null values from input dataset - (except key fields) during write operation. Default is false. Type: boolean (or Expression with - resultType boolean). - :type ignore_null_values: any - :param alternate_key_name: The logical name of the alternate key which will be used when - upserting records. Type: string (or Expression with resultType string). - :type alternate_key_name: any - """ - - _validation = { - 'type': {'required': True}, - 'write_behavior': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, - 'ignore_null_values': {'key': 'ignoreNullValues', 'type': 'object'}, - 'alternate_key_name': {'key': 'alternateKeyName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DynamicsCrmSink, self).__init__(**kwargs) - self.type = 'DynamicsCrmSink' # type: str - self.write_behavior = kwargs['write_behavior'] - self.ignore_null_values = kwargs.get('ignore_null_values', None) - self.alternate_key_name = kwargs.get('alternate_key_name', None) - - -class DynamicsCrmSource(CopySource): - """A copy activity Dynamics CRM source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: FetchXML is a proprietary query language that is used in Microsoft Dynamics CRM - (online & on-premises). Type: string (or Expression with resultType string). - :type query: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DynamicsCrmSource, self).__init__(**kwargs) - self.type = 'DynamicsCrmSource' # type: str - self.query = kwargs.get('query', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class DynamicsEntityDataset(Dataset): - """The Dynamics entity dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param entity_name: The logical name of the entity. Type: string (or Expression with resultType - string). - :type entity_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'entity_name': {'key': 'typeProperties.entityName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DynamicsEntityDataset, self).__init__(**kwargs) - self.type = 'DynamicsEntity' # type: str - self.entity_name = kwargs.get('entity_name', None) - - -class DynamicsLinkedService(LinkedService): - """Dynamics linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param deployment_type: Required. The deployment type of the Dynamics instance. 'Online' for - Dynamics Online and 'OnPremisesWithIfd' for Dynamics on-premises with Ifd. Type: string (or - Expression with resultType string). Possible values include: "Online", "OnPremisesWithIfd". - :type deployment_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsDeploymentType - :param host_name: The host name of the on-premises Dynamics server. The property is required - for on-prem and not allowed for online. Type: string (or Expression with resultType string). - :type host_name: any - :param port: The port of on-premises Dynamics server. The property is required for on-prem and - not allowed for online. Default is 443. Type: integer (or Expression with resultType integer), - minimum: 0. - :type port: any - :param service_uri: The URL to the Microsoft Dynamics server. The property is required for - on-line and not allowed for on-prem. Type: string (or Expression with resultType string). - :type service_uri: any - :param organization_name: The organization name of the Dynamics instance. The property is - required for on-prem and required for online when there are more than one Dynamics instances - associated with the user. Type: string (or Expression with resultType string). - :type organization_name: any - :param authentication_type: Required. The authentication type to connect to Dynamics server. - 'Office365' for online scenario, 'Ifd' for on-premises with Ifd scenario, 'AADServicePrincipal' - for Server-To-Server authentication in online scenario. Type: string (or Expression with - resultType string). Possible values include: "Office365", "Ifd", "AADServicePrincipal". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsAuthenticationType - :param username: User name to access the Dynamics instance. Type: string (or Expression with - resultType string). - :type username: any - :param password: Password to access the Dynamics instance. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param service_principal_id: The client ID of the application in Azure Active Directory used - for Server-To-Server authentication. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_credential_type: The service principal credential type to use in - Server-To-Server authentication. 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' - for certificate. Type: string (or Expression with resultType string). Possible values include: - "ServicePrincipalKey", "ServicePrincipalCert". - :type service_principal_credential_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsServicePrincipalCredentialType - :param service_principal_credential: The credential of the service principal object in Azure - Active Directory. If servicePrincipalCredentialType is 'ServicePrincipalKey', - servicePrincipalCredential can be SecureString or AzureKeyVaultSecretReference. If - servicePrincipalCredentialType is 'ServicePrincipalCert', servicePrincipalCredential can only - be AzureKeyVaultSecretReference. - :type service_principal_credential: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'deployment_type': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'str'}, - 'host_name': {'key': 'typeProperties.hostName', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'service_uri': {'key': 'typeProperties.serviceUri', 'type': 'object'}, - 'organization_name': {'key': 'typeProperties.organizationName', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_credential_type': {'key': 'typeProperties.servicePrincipalCredentialType', 'type': 'str'}, - 'service_principal_credential': {'key': 'typeProperties.servicePrincipalCredential', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DynamicsLinkedService, self).__init__(**kwargs) - self.type = 'Dynamics' # type: str - self.deployment_type = kwargs['deployment_type'] - self.host_name = kwargs.get('host_name', None) - self.port = kwargs.get('port', None) - self.service_uri = kwargs.get('service_uri', None) - self.organization_name = kwargs.get('organization_name', None) - self.authentication_type = kwargs['authentication_type'] - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.service_principal_id = kwargs.get('service_principal_id', None) - self.service_principal_credential_type = kwargs.get('service_principal_credential_type', None) - self.service_principal_credential = kwargs.get('service_principal_credential', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class DynamicsSink(CopySink): - """A copy activity Dynamics sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param write_behavior: Required. The write behavior for the operation. Possible values include: - "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsSinkWriteBehavior - :param ignore_null_values: The flag indicating whether ignore null values from input dataset - (except key fields) during write operation. Default is false. Type: boolean (or Expression with - resultType boolean). - :type ignore_null_values: any - :param alternate_key_name: The logical name of the alternate key which will be used when - upserting records. Type: string (or Expression with resultType string). - :type alternate_key_name: any - """ - - _validation = { - 'type': {'required': True}, - 'write_behavior': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, - 'ignore_null_values': {'key': 'ignoreNullValues', 'type': 'object'}, - 'alternate_key_name': {'key': 'alternateKeyName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DynamicsSink, self).__init__(**kwargs) - self.type = 'DynamicsSink' # type: str - self.write_behavior = kwargs['write_behavior'] - self.ignore_null_values = kwargs.get('ignore_null_values', None) - self.alternate_key_name = kwargs.get('alternate_key_name', None) - - -class DynamicsSource(CopySource): - """A copy activity Dynamics source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: FetchXML is a proprietary query language that is used in Microsoft Dynamics - (online & on-premises). Type: string (or Expression with resultType string). - :type query: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(DynamicsSource, self).__init__(**kwargs) - self.type = 'DynamicsSource' # type: str - self.query = kwargs.get('query', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class EloquaLinkedService(LinkedService): - """Eloqua server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param endpoint: Required. The endpoint of the Eloqua server. (i.e. eloqua.example.com). - :type endpoint: any - :param username: Required. The site name and user name of your Eloqua account in the form: - sitename/username. (i.e. Eloqua/Alice). - :type username: any - :param password: The password corresponding to the user name. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'endpoint': {'required': True}, - 'username': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(EloquaLinkedService, self).__init__(**kwargs) - self.type = 'Eloqua' # type: str - self.endpoint = kwargs['endpoint'] - self.username = kwargs['username'] - self.password = kwargs.get('password', None) - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.use_host_verification = kwargs.get('use_host_verification', None) - self.use_peer_verification = kwargs.get('use_peer_verification', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class EloquaObjectDataset(Dataset): - """Eloqua server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(EloquaObjectDataset, self).__init__(**kwargs) - self.type = 'EloquaObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class EloquaSource(TabularSource): - """A copy activity Eloqua server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(EloquaSource, self).__init__(**kwargs) - self.type = 'EloquaSource' # type: str - self.query = kwargs.get('query', None) - - -class EncryptionDetails(msrest.serialization.Model): - """Details of the encryption associated with the workspace. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar double_encryption_enabled: Double Encryption enabled. - :vartype double_encryption_enabled: bool - :param cmk: Customer Managed Key Details. - :type cmk: ~azure.synapse.artifacts.v2020_12_01.models.CustomerManagedKeyDetails - """ - - _validation = { - 'double_encryption_enabled': {'readonly': True}, - } - - _attribute_map = { - 'double_encryption_enabled': {'key': 'doubleEncryptionEnabled', 'type': 'bool'}, - 'cmk': {'key': 'cmk', 'type': 'CustomerManagedKeyDetails'}, - } - - def __init__( - self, - **kwargs - ): - super(EncryptionDetails, self).__init__(**kwargs) - self.double_encryption_enabled = None - self.cmk = kwargs.get('cmk', None) - - -class EntityReference(msrest.serialization.Model): - """The entity reference. - - :param type: The type of this referenced entity. Possible values include: - "IntegrationRuntimeReference", "LinkedServiceReference". - :type type: str or - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeEntityReferenceType - :param reference_name: The name of this referenced entity. - :type reference_name: str - """ - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(EntityReference, self).__init__(**kwargs) - self.type = kwargs.get('type', None) - self.reference_name = kwargs.get('reference_name', None) - - -class ErrorAdditionalInfo(msrest.serialization.Model): - """The resource management error additional info. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar type: The additional info type. - :vartype type: str - :ivar info: The additional info. - :vartype info: any - """ - - _validation = { - 'type': {'readonly': True}, - 'info': {'readonly': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'info': {'key': 'info', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ErrorAdditionalInfo, self).__init__(**kwargs) - self.type = None - self.info = None - - -class ErrorContract(msrest.serialization.Model): - """Contains details when the response code indicates an error. - - :param error: The error details. - :type error: ~azure.synapse.artifacts.v2020_12_01.models.ErrorResponse - """ - - _attribute_map = { - 'error': {'key': 'error', 'type': 'ErrorResponse'}, - } - - def __init__( - self, - **kwargs - ): - super(ErrorContract, self).__init__(**kwargs) - self.error = kwargs.get('error', None) - - -class ErrorResponse(msrest.serialization.Model): - """Common error response for all Azure Resource Manager APIs to return error details for failed operations. (This also follows the OData error response format.). - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar code: The error code. - :vartype code: str - :ivar message: The error message. - :vartype message: str - :ivar target: The error target. - :vartype target: str - :ivar details: The error details. - :vartype details: list[~azure.synapse.artifacts.v2020_12_01.models.ErrorResponse] - :ivar additional_info: The error additional info. - :vartype additional_info: list[~azure.synapse.artifacts.v2020_12_01.models.ErrorAdditionalInfo] - """ - - _validation = { - 'code': {'readonly': True}, - 'message': {'readonly': True}, - 'target': {'readonly': True}, - 'details': {'readonly': True}, - 'additional_info': {'readonly': True}, - } - - _attribute_map = { - 'code': {'key': 'code', 'type': 'str'}, - 'message': {'key': 'message', 'type': 'str'}, - 'target': {'key': 'target', 'type': 'str'}, - 'details': {'key': 'details', 'type': '[ErrorResponse]'}, - 'additional_info': {'key': 'additionalInfo', 'type': '[ErrorAdditionalInfo]'}, - } - - def __init__( - self, - **kwargs - ): - super(ErrorResponse, self).__init__(**kwargs) - self.code = None - self.message = None - self.target = None - self.details = None - self.additional_info = None - - -class EvaluateDataFlowExpressionRequest(msrest.serialization.Model): - """Request body structure for data flow expression preview. - - :param session_id: The ID of data flow debug session. - :type session_id: str - :param data_flow_name: The data flow which contains the debug session. - :type data_flow_name: str - :param stream_name: The output stream name. - :type stream_name: str - :param row_limits: The row limit for preview request. - :type row_limits: int - :param expression: The expression for preview. - :type expression: str - """ - - _attribute_map = { - 'session_id': {'key': 'sessionId', 'type': 'str'}, - 'data_flow_name': {'key': 'dataFlowName', 'type': 'str'}, - 'stream_name': {'key': 'streamName', 'type': 'str'}, - 'row_limits': {'key': 'rowLimits', 'type': 'int'}, - 'expression': {'key': 'expression', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(EvaluateDataFlowExpressionRequest, self).__init__(**kwargs) - self.session_id = kwargs.get('session_id', None) - self.data_flow_name = kwargs.get('data_flow_name', None) - self.stream_name = kwargs.get('stream_name', None) - self.row_limits = kwargs.get('row_limits', None) - self.expression = kwargs.get('expression', None) - - -class ExcelDataset(Dataset): - """Excel dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param location: The location of the excel storage. - :type location: ~azure.synapse.artifacts.v2020_12_01.models.DatasetLocation - :param sheet_name: The sheet of excel file. Type: string (or Expression with resultType - string). - :type sheet_name: any - :param range: The partial data of one sheet. Type: string (or Expression with resultType - string). - :type range: any - :param first_row_as_header: When used as input, treat the first row of data as headers. When - used as output,write the headers into the output as the first row of data. The default value is - false. Type: boolean (or Expression with resultType boolean). - :type first_row_as_header: any - :param compression: The data compression method used for the json dataset. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - :param null_value: The null value string. Type: string (or Expression with resultType string). - :type null_value: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'location': {'key': 'typeProperties.location', 'type': 'DatasetLocation'}, - 'sheet_name': {'key': 'typeProperties.sheetName', 'type': 'object'}, - 'range': {'key': 'typeProperties.range', 'type': 'object'}, - 'first_row_as_header': {'key': 'typeProperties.firstRowAsHeader', 'type': 'object'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - 'null_value': {'key': 'typeProperties.nullValue', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ExcelDataset, self).__init__(**kwargs) - self.type = 'Excel' # type: str - self.location = kwargs.get('location', None) - self.sheet_name = kwargs.get('sheet_name', None) - self.range = kwargs.get('range', None) - self.first_row_as_header = kwargs.get('first_row_as_header', None) - self.compression = kwargs.get('compression', None) - self.null_value = kwargs.get('null_value', None) - - -class ExcelSource(CopySource): - """A copy activity excel source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Excel store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ExcelSource, self).__init__(**kwargs) - self.type = 'ExcelSource' # type: str - self.store_settings = kwargs.get('store_settings', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class ExecuteDataFlowActivity(ExecutionActivity): - """Execute data flow activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param dataflow: Required. Data flow reference. - :type dataflow: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowReference - :param staging: Staging info for execute data flow activity. - :type staging: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowStagingInfo - :param integration_runtime: The integration runtime reference. - :type integration_runtime: - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param compute: Compute properties for data flow activity. - :type compute: - ~azure.synapse.artifacts.v2020_12_01.models.ExecuteDataFlowActivityTypePropertiesCompute - :param trace_level: Trace level setting used for data flow monitoring output. Supported values - are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string). - :type trace_level: any - :param continue_on_error: Continue on error setting used for data flow execution. Enables - processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean). - :type continue_on_error: any - :param run_concurrently: Concurrent run setting used for data flow execution. Allows sinks with - the same save order to be processed concurrently. Type: boolean (or Expression with resultType - boolean). - :type run_concurrently: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'dataflow': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'dataflow': {'key': 'typeProperties.dataflow', 'type': 'DataFlowReference'}, - 'staging': {'key': 'typeProperties.staging', 'type': 'DataFlowStagingInfo'}, - 'integration_runtime': {'key': 'typeProperties.integrationRuntime', 'type': 'IntegrationRuntimeReference'}, - 'compute': {'key': 'typeProperties.compute', 'type': 'ExecuteDataFlowActivityTypePropertiesCompute'}, - 'trace_level': {'key': 'typeProperties.traceLevel', 'type': 'object'}, - 'continue_on_error': {'key': 'typeProperties.continueOnError', 'type': 'object'}, - 'run_concurrently': {'key': 'typeProperties.runConcurrently', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ExecuteDataFlowActivity, self).__init__(**kwargs) - self.type = 'ExecuteDataFlow' # type: str - self.dataflow = kwargs['dataflow'] - self.staging = kwargs.get('staging', None) - self.integration_runtime = kwargs.get('integration_runtime', None) - self.compute = kwargs.get('compute', None) - self.trace_level = kwargs.get('trace_level', None) - self.continue_on_error = kwargs.get('continue_on_error', None) - self.run_concurrently = kwargs.get('run_concurrently', None) - - -class ExecuteDataFlowActivityTypePropertiesCompute(msrest.serialization.Model): - """Compute properties for data flow activity. - - :param compute_type: Compute type of the cluster which will execute data flow job. Possible - values include: "General", "MemoryOptimized", "ComputeOptimized". - :type compute_type: str or ~azure.synapse.artifacts.v2020_12_01.models.DataFlowComputeType - :param core_count: Core count of the cluster which will execute data flow job. Supported values - are: 8, 16, 32, 48, 80, 144 and 272. - :type core_count: int - """ - - _attribute_map = { - 'compute_type': {'key': 'computeType', 'type': 'str'}, - 'core_count': {'key': 'coreCount', 'type': 'int'}, - } - - def __init__( - self, - **kwargs - ): - super(ExecuteDataFlowActivityTypePropertiesCompute, self).__init__(**kwargs) - self.compute_type = kwargs.get('compute_type', None) - self.core_count = kwargs.get('core_count', None) - - -class ExecutePipelineActivity(ControlActivity): - """Execute pipeline activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param pipeline: Required. Pipeline reference. - :type pipeline: ~azure.synapse.artifacts.v2020_12_01.models.PipelineReference - :param parameters: Pipeline parameters. - :type parameters: dict[str, any] - :param wait_on_completion: Defines whether activity execution will wait for the dependent - pipeline execution to finish. Default is false. - :type wait_on_completion: bool - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'pipeline': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'pipeline': {'key': 'typeProperties.pipeline', 'type': 'PipelineReference'}, - 'parameters': {'key': 'typeProperties.parameters', 'type': '{object}'}, - 'wait_on_completion': {'key': 'typeProperties.waitOnCompletion', 'type': 'bool'}, - } - - def __init__( - self, - **kwargs - ): - super(ExecutePipelineActivity, self).__init__(**kwargs) - self.type = 'ExecutePipeline' # type: str - self.pipeline = kwargs['pipeline'] - self.parameters = kwargs.get('parameters', None) - self.wait_on_completion = kwargs.get('wait_on_completion', None) - - -class ExecuteSSISPackageActivity(ExecutionActivity): - """Execute SSIS package activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param package_location: Required. SSIS package location. - :type package_location: ~azure.synapse.artifacts.v2020_12_01.models.SSISPackageLocation - :param runtime: Specifies the runtime to execute SSIS package. The value should be "x86" or - "x64". Type: string (or Expression with resultType string). - :type runtime: any - :param logging_level: The logging level of SSIS package execution. Type: string (or Expression - with resultType string). - :type logging_level: any - :param environment_path: The environment path to execute the SSIS package. Type: string (or - Expression with resultType string). - :type environment_path: any - :param execution_credential: The package execution credential. - :type execution_credential: ~azure.synapse.artifacts.v2020_12_01.models.SSISExecutionCredential - :param connect_via: Required. The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param project_parameters: The project level parameters to execute the SSIS package. - :type project_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.SSISExecutionParameter] - :param package_parameters: The package level parameters to execute the SSIS package. - :type package_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.SSISExecutionParameter] - :param project_connection_managers: The project level connection managers to execute the SSIS - package. - :type project_connection_managers: dict[str, dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.SSISExecutionParameter]] - :param package_connection_managers: The package level connection managers to execute the SSIS - package. - :type package_connection_managers: dict[str, dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.SSISExecutionParameter]] - :param property_overrides: The property overrides to execute the SSIS package. - :type property_overrides: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.SSISPropertyOverride] - :param log_location: SSIS package execution log location. - :type log_location: ~azure.synapse.artifacts.v2020_12_01.models.SSISLogLocation - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'package_location': {'required': True}, - 'connect_via': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'package_location': {'key': 'typeProperties.packageLocation', 'type': 'SSISPackageLocation'}, - 'runtime': {'key': 'typeProperties.runtime', 'type': 'object'}, - 'logging_level': {'key': 'typeProperties.loggingLevel', 'type': 'object'}, - 'environment_path': {'key': 'typeProperties.environmentPath', 'type': 'object'}, - 'execution_credential': {'key': 'typeProperties.executionCredential', 'type': 'SSISExecutionCredential'}, - 'connect_via': {'key': 'typeProperties.connectVia', 'type': 'IntegrationRuntimeReference'}, - 'project_parameters': {'key': 'typeProperties.projectParameters', 'type': '{SSISExecutionParameter}'}, - 'package_parameters': {'key': 'typeProperties.packageParameters', 'type': '{SSISExecutionParameter}'}, - 'project_connection_managers': {'key': 'typeProperties.projectConnectionManagers', 'type': '{{SSISExecutionParameter}}'}, - 'package_connection_managers': {'key': 'typeProperties.packageConnectionManagers', 'type': '{{SSISExecutionParameter}}'}, - 'property_overrides': {'key': 'typeProperties.propertyOverrides', 'type': '{SSISPropertyOverride}'}, - 'log_location': {'key': 'typeProperties.logLocation', 'type': 'SSISLogLocation'}, - } - - def __init__( - self, - **kwargs - ): - super(ExecuteSSISPackageActivity, self).__init__(**kwargs) - self.type = 'ExecuteSSISPackage' # type: str - self.package_location = kwargs['package_location'] - self.runtime = kwargs.get('runtime', None) - self.logging_level = kwargs.get('logging_level', None) - self.environment_path = kwargs.get('environment_path', None) - self.execution_credential = kwargs.get('execution_credential', None) - self.connect_via = kwargs['connect_via'] - self.project_parameters = kwargs.get('project_parameters', None) - self.package_parameters = kwargs.get('package_parameters', None) - self.project_connection_managers = kwargs.get('project_connection_managers', None) - self.package_connection_managers = kwargs.get('package_connection_managers', None) - self.property_overrides = kwargs.get('property_overrides', None) - self.log_location = kwargs.get('log_location', None) - - -class ExposureControlRequest(msrest.serialization.Model): - """The exposure control request. - - :param feature_name: The feature name. - :type feature_name: str - :param feature_type: The feature type. - :type feature_type: str - """ - - _attribute_map = { - 'feature_name': {'key': 'featureName', 'type': 'str'}, - 'feature_type': {'key': 'featureType', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(ExposureControlRequest, self).__init__(**kwargs) - self.feature_name = kwargs.get('feature_name', None) - self.feature_type = kwargs.get('feature_type', None) - - -class ExposureControlResponse(msrest.serialization.Model): - """The exposure control response. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar feature_name: The feature name. - :vartype feature_name: str - :ivar value: The feature value. - :vartype value: str - """ - - _validation = { - 'feature_name': {'readonly': True}, - 'value': {'readonly': True}, - } - - _attribute_map = { - 'feature_name': {'key': 'featureName', 'type': 'str'}, - 'value': {'key': 'value', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(ExposureControlResponse, self).__init__(**kwargs) - self.feature_name = None - self.value = None - - -class Expression(msrest.serialization.Model): - """Azure Synapse expression definition. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Expression type. Possible values include: "Expression". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.ExpressionType - :param value: Required. Expression value. - :type value: str - """ - - _validation = { - 'type': {'required': True}, - 'value': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'value': {'key': 'value', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(Expression, self).__init__(**kwargs) - self.type = kwargs['type'] - self.value = kwargs['value'] - - -class FileServerLinkedService(LinkedService): - """File system linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. Host name of the server. Type: string (or Expression with resultType - string). - :type host: any - :param user_id: User ID to logon the server. Type: string (or Expression with resultType - string). - :type user_id: any - :param password: Password to logon the server. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'user_id': {'key': 'typeProperties.userId', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(FileServerLinkedService, self).__init__(**kwargs) - self.type = 'FileServer' # type: str - self.host = kwargs['host'] - self.user_id = kwargs.get('user_id', None) - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class FileServerLocation(DatasetLocation): - """The location of file server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(FileServerLocation, self).__init__(**kwargs) - self.type = 'FileServerLocation' # type: str - - -class FileServerReadSettings(StoreReadSettings): - """File server read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: FileServer wildcardFolderPath. Type: string (or Expression with - resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: FileServer wildcardFileName. Type: string (or Expression with - resultType string). - :type wildcard_file_name: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - :param file_filter: Specify a filter to be used to select a subset of files in the folderPath - rather than all files. Type: string (or Expression with resultType string). - :type file_filter: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - 'modified_datetime_start': {'key': 'modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'modifiedDatetimeEnd', 'type': 'object'}, - 'file_filter': {'key': 'fileFilter', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(FileServerReadSettings, self).__init__(**kwargs) - self.type = 'FileServerReadSettings' # type: str - self.recursive = kwargs.get('recursive', None) - self.wildcard_folder_path = kwargs.get('wildcard_folder_path', None) - self.wildcard_file_name = kwargs.get('wildcard_file_name', None) - self.file_list_path = kwargs.get('file_list_path', None) - self.enable_partition_discovery = kwargs.get('enable_partition_discovery', None) - self.partition_root_path = kwargs.get('partition_root_path', None) - self.delete_files_after_completion = kwargs.get('delete_files_after_completion', None) - self.modified_datetime_start = kwargs.get('modified_datetime_start', None) - self.modified_datetime_end = kwargs.get('modified_datetime_end', None) - self.file_filter = kwargs.get('file_filter', None) - - -class FileServerWriteSettings(StoreWriteSettings): - """File server write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(FileServerWriteSettings, self).__init__(**kwargs) - self.type = 'FileServerWriteSettings' # type: str - - -class FileShareDataset(Dataset): - """An on-premises file system dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param folder_path: The path of the on-premises file system. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: The name of the on-premises file system. Type: string (or Expression with - resultType string). - :type file_name: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - :param format: The format of the files. - :type format: ~azure.synapse.artifacts.v2020_12_01.models.DatasetStorageFormat - :param file_filter: Specify a filter to be used to select a subset of files in the folderPath - rather than all files. Type: string (or Expression with resultType string). - :type file_filter: any - :param compression: The data compression method used for the file system. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, - 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, - 'modified_datetime_start': {'key': 'typeProperties.modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'typeProperties.modifiedDatetimeEnd', 'type': 'object'}, - 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, - 'file_filter': {'key': 'typeProperties.fileFilter', 'type': 'object'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - } - - def __init__( - self, - **kwargs - ): - super(FileShareDataset, self).__init__(**kwargs) - self.type = 'FileShare' # type: str - self.folder_path = kwargs.get('folder_path', None) - self.file_name = kwargs.get('file_name', None) - self.modified_datetime_start = kwargs.get('modified_datetime_start', None) - self.modified_datetime_end = kwargs.get('modified_datetime_end', None) - self.format = kwargs.get('format', None) - self.file_filter = kwargs.get('file_filter', None) - self.compression = kwargs.get('compression', None) - - -class FileSystemSink(CopySink): - """A copy activity file system sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(FileSystemSink, self).__init__(**kwargs) - self.type = 'FileSystemSink' # type: str - self.copy_behavior = kwargs.get('copy_behavior', None) - - -class FileSystemSource(CopySource): - """A copy activity file system source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(FileSystemSource, self).__init__(**kwargs) - self.type = 'FileSystemSource' # type: str - self.recursive = kwargs.get('recursive', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class FilterActivity(ControlActivity): - """Filter and return results from input array based on the conditions. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param items: Required. Input array on which filter should be applied. - :type items: ~azure.synapse.artifacts.v2020_12_01.models.Expression - :param condition: Required. Condition to be used for filtering the input. - :type condition: ~azure.synapse.artifacts.v2020_12_01.models.Expression - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'items': {'required': True}, - 'condition': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'items': {'key': 'typeProperties.items', 'type': 'Expression'}, - 'condition': {'key': 'typeProperties.condition', 'type': 'Expression'}, - } - - def __init__( - self, - **kwargs - ): - super(FilterActivity, self).__init__(**kwargs) - self.type = 'Filter' # type: str - self.items = kwargs['items'] - self.condition = kwargs['condition'] - - -class ForEachActivity(ControlActivity): - """This activity is used for iterating over a collection and execute given activities. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param is_sequential: Should the loop be executed in sequence or in parallel (max 50). - :type is_sequential: bool - :param batch_count: Batch count to be used for controlling the number of parallel execution - (when isSequential is set to false). - :type batch_count: int - :param items: Required. Collection to iterate. - :type items: ~azure.synapse.artifacts.v2020_12_01.models.Expression - :param activities: Required. List of activities to execute . - :type activities: list[~azure.synapse.artifacts.v2020_12_01.models.Activity] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'batch_count': {'maximum': 50}, - 'items': {'required': True}, - 'activities': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'is_sequential': {'key': 'typeProperties.isSequential', 'type': 'bool'}, - 'batch_count': {'key': 'typeProperties.batchCount', 'type': 'int'}, - 'items': {'key': 'typeProperties.items', 'type': 'Expression'}, - 'activities': {'key': 'typeProperties.activities', 'type': '[Activity]'}, - } - - def __init__( - self, - **kwargs - ): - super(ForEachActivity, self).__init__(**kwargs) - self.type = 'ForEach' # type: str - self.is_sequential = kwargs.get('is_sequential', None) - self.batch_count = kwargs.get('batch_count', None) - self.items = kwargs['items'] - self.activities = kwargs['activities'] - - -class FtpReadSettings(StoreReadSettings): - """Ftp read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: Ftp wildcardFolderPath. Type: string (or Expression with - resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: Ftp wildcardFileName. Type: string (or Expression with resultType - string). - :type wildcard_file_name: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param use_binary_transfer: Specify whether to use binary transfer mode for FTP stores. - :type use_binary_transfer: bool - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'use_binary_transfer': {'key': 'useBinaryTransfer', 'type': 'bool'}, - } - - def __init__( - self, - **kwargs - ): - super(FtpReadSettings, self).__init__(**kwargs) - self.type = 'FtpReadSettings' # type: str - self.recursive = kwargs.get('recursive', None) - self.wildcard_folder_path = kwargs.get('wildcard_folder_path', None) - self.wildcard_file_name = kwargs.get('wildcard_file_name', None) - self.enable_partition_discovery = kwargs.get('enable_partition_discovery', None) - self.partition_root_path = kwargs.get('partition_root_path', None) - self.delete_files_after_completion = kwargs.get('delete_files_after_completion', None) - self.file_list_path = kwargs.get('file_list_path', None) - self.use_binary_transfer = kwargs.get('use_binary_transfer', None) - - -class FtpServerLinkedService(LinkedService): - """A FTP server Linked Service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. Host name of the FTP server. Type: string (or Expression with resultType - string). - :type host: any - :param port: The TCP port number that the FTP server uses to listen for client connections. - Default value is 21. Type: integer (or Expression with resultType integer), minimum: 0. - :type port: any - :param authentication_type: The authentication type to be used to connect to the FTP server. - Possible values include: "Basic", "Anonymous". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.FtpAuthenticationType - :param user_name: Username to logon the FTP server. Type: string (or Expression with resultType - string). - :type user_name: any - :param password: Password to logon the FTP server. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - :param enable_ssl: If true, connect to the FTP server over SSL/TLS channel. Default value is - true. Type: boolean (or Expression with resultType boolean). - :type enable_ssl: any - :param enable_server_certificate_validation: If true, validate the FTP server SSL certificate - when connect over SSL/TLS channel. Default value is true. Type: boolean (or Expression with - resultType boolean). - :type enable_server_certificate_validation: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, - 'enable_server_certificate_validation': {'key': 'typeProperties.enableServerCertificateValidation', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(FtpServerLinkedService, self).__init__(**kwargs) - self.type = 'FtpServer' # type: str - self.host = kwargs['host'] - self.port = kwargs.get('port', None) - self.authentication_type = kwargs.get('authentication_type', None) - self.user_name = kwargs.get('user_name', None) - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - self.enable_ssl = kwargs.get('enable_ssl', None) - self.enable_server_certificate_validation = kwargs.get('enable_server_certificate_validation', None) - - -class FtpServerLocation(DatasetLocation): - """The location of ftp server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(FtpServerLocation, self).__init__(**kwargs) - self.type = 'FtpServerLocation' # type: str - - -class GetMetadataActivity(ExecutionActivity): - """Activity to get metadata of dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param dataset: Required. GetMetadata activity dataset reference. - :type dataset: ~azure.synapse.artifacts.v2020_12_01.models.DatasetReference - :param field_list: Fields of metadata to get from dataset. - :type field_list: list[any] - :param store_settings: GetMetadata activity store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - :param format_settings: GetMetadata activity format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.FormatReadSettings - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'dataset': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'dataset': {'key': 'typeProperties.dataset', 'type': 'DatasetReference'}, - 'field_list': {'key': 'typeProperties.fieldList', 'type': '[object]'}, - 'store_settings': {'key': 'typeProperties.storeSettings', 'type': 'StoreReadSettings'}, - 'format_settings': {'key': 'typeProperties.formatSettings', 'type': 'FormatReadSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(GetMetadataActivity, self).__init__(**kwargs) - self.type = 'GetMetadata' # type: str - self.dataset = kwargs['dataset'] - self.field_list = kwargs.get('field_list', None) - self.store_settings = kwargs.get('store_settings', None) - self.format_settings = kwargs.get('format_settings', None) - - -class GetSsisObjectMetadataRequest(msrest.serialization.Model): - """The request payload of get SSIS object metadata. - - :param metadata_path: Metadata path. - :type metadata_path: str - """ - - _attribute_map = { - 'metadata_path': {'key': 'metadataPath', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(GetSsisObjectMetadataRequest, self).__init__(**kwargs) - self.metadata_path = kwargs.get('metadata_path', None) - - -class GitHubAccessTokenRequest(msrest.serialization.Model): - """GitHubAccessTokenRequest. - - All required parameters must be populated in order to send to Azure. - - :param git_hub_client_id: Required. The GitHub Client Id. - :type git_hub_client_id: str - :param git_hub_access_code: Required. The GitHub Access code. - :type git_hub_access_code: str - :param git_hub_access_token_base_url: Required. The GitHub access token base URL. - :type git_hub_access_token_base_url: str - """ - - _validation = { - 'git_hub_client_id': {'required': True}, - 'git_hub_access_code': {'required': True}, - 'git_hub_access_token_base_url': {'required': True}, - } - - _attribute_map = { - 'git_hub_client_id': {'key': 'gitHubClientId', 'type': 'str'}, - 'git_hub_access_code': {'key': 'gitHubAccessCode', 'type': 'str'}, - 'git_hub_access_token_base_url': {'key': 'gitHubAccessTokenBaseUrl', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(GitHubAccessTokenRequest, self).__init__(**kwargs) - self.git_hub_client_id = kwargs['git_hub_client_id'] - self.git_hub_access_code = kwargs['git_hub_access_code'] - self.git_hub_access_token_base_url = kwargs['git_hub_access_token_base_url'] - - -class GitHubAccessTokenResponse(msrest.serialization.Model): - """GitHubAccessTokenResponse. - - :param git_hub_access_token: - :type git_hub_access_token: str - """ - - _attribute_map = { - 'git_hub_access_token': {'key': 'gitHubAccessToken', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(GitHubAccessTokenResponse, self).__init__(**kwargs) - self.git_hub_access_token = kwargs.get('git_hub_access_token', None) - - -class GitHubClientSecret(msrest.serialization.Model): - """Client secret information for factory's bring your own app repository configuration. - - :param byoa_secret_akv_url: Bring your own app client secret AKV URL. - :type byoa_secret_akv_url: str - :param byoa_secret_name: Bring your own app client secret name in AKV. - :type byoa_secret_name: str - """ - - _attribute_map = { - 'byoa_secret_akv_url': {'key': 'byoaSecretAkvUrl', 'type': 'str'}, - 'byoa_secret_name': {'key': 'byoaSecretName', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(GitHubClientSecret, self).__init__(**kwargs) - self.byoa_secret_akv_url = kwargs.get('byoa_secret_akv_url', None) - self.byoa_secret_name = kwargs.get('byoa_secret_name', None) - - -class GoogleAdWordsLinkedService(LinkedService): - """Google AdWords service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param client_customer_id: Required. The Client customer ID of the AdWords account that you - want to fetch report data for. - :type client_customer_id: any - :param developer_token: Required. The developer token associated with the manager account that - you use to grant access to the AdWords API. - :type developer_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param authentication_type: Required. The OAuth 2.0 authentication mechanism used for - authentication. ServiceAuthentication can only be used on self-hosted IR. Possible values - include: "ServiceAuthentication", "UserAuthentication". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.GoogleAdWordsAuthenticationType - :param refresh_token: The refresh token obtained from Google for authorizing access to AdWords - for UserAuthentication. - :type refresh_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param client_id: The client id of the google application used to acquire the refresh token. - Type: string (or Expression with resultType string). - :type client_id: any - :param client_secret: The client secret of the google application used to acquire the refresh - token. - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param email: The service account email ID that is used for ServiceAuthentication and can only - be used on self-hosted IR. - :type email: any - :param key_file_path: The full path to the .p12 key file that is used to authenticate the - service account email address and can only be used on self-hosted IR. - :type key_file_path: any - :param trusted_cert_path: The full path of the .pem file containing trusted CA certificates for - verifying the server when connecting over SSL. This property can only be set when using SSL on - self-hosted IR. The default value is the cacerts.pem file installed with the IR. - :type trusted_cert_path: any - :param use_system_trust_store: Specifies whether to use a CA certificate from the system trust - store or from a specified PEM file. The default value is false. - :type use_system_trust_store: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'client_customer_id': {'required': True}, - 'developer_token': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'client_customer_id': {'key': 'typeProperties.clientCustomerID', 'type': 'object'}, - 'developer_token': {'key': 'typeProperties.developerToken', 'type': 'SecretBase'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'refresh_token': {'key': 'typeProperties.refreshToken', 'type': 'SecretBase'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'email': {'key': 'typeProperties.email', 'type': 'object'}, - 'key_file_path': {'key': 'typeProperties.keyFilePath', 'type': 'object'}, - 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, - 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(GoogleAdWordsLinkedService, self).__init__(**kwargs) - self.type = 'GoogleAdWords' # type: str - self.client_customer_id = kwargs['client_customer_id'] - self.developer_token = kwargs['developer_token'] - self.authentication_type = kwargs['authentication_type'] - self.refresh_token = kwargs.get('refresh_token', None) - self.client_id = kwargs.get('client_id', None) - self.client_secret = kwargs.get('client_secret', None) - self.email = kwargs.get('email', None) - self.key_file_path = kwargs.get('key_file_path', None) - self.trusted_cert_path = kwargs.get('trusted_cert_path', None) - self.use_system_trust_store = kwargs.get('use_system_trust_store', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class GoogleAdWordsObjectDataset(Dataset): - """Google AdWords service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(GoogleAdWordsObjectDataset, self).__init__(**kwargs) - self.type = 'GoogleAdWordsObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class GoogleAdWordsSource(TabularSource): - """A copy activity Google AdWords service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(GoogleAdWordsSource, self).__init__(**kwargs) - self.type = 'GoogleAdWordsSource' # type: str - self.query = kwargs.get('query', None) - - -class GoogleBigQueryLinkedService(LinkedService): - """Google BigQuery service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param project: Required. The default BigQuery project to query against. - :type project: any - :param additional_projects: A comma-separated list of public BigQuery projects to access. - :type additional_projects: any - :param request_google_drive_scope: Whether to request access to Google Drive. Allowing Google - Drive access enables support for federated tables that combine BigQuery data with data from - Google Drive. The default value is false. - :type request_google_drive_scope: any - :param authentication_type: Required. The OAuth 2.0 authentication mechanism used for - authentication. ServiceAuthentication can only be used on self-hosted IR. Possible values - include: "ServiceAuthentication", "UserAuthentication". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.GoogleBigQueryAuthenticationType - :param refresh_token: The refresh token obtained from Google for authorizing access to BigQuery - for UserAuthentication. - :type refresh_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param client_id: The client id of the google application used to acquire the refresh token. - Type: string (or Expression with resultType string). - :type client_id: any - :param client_secret: The client secret of the google application used to acquire the refresh - token. - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param email: The service account email ID that is used for ServiceAuthentication and can only - be used on self-hosted IR. - :type email: any - :param key_file_path: The full path to the .p12 key file that is used to authenticate the - service account email address and can only be used on self-hosted IR. - :type key_file_path: any - :param trusted_cert_path: The full path of the .pem file containing trusted CA certificates for - verifying the server when connecting over SSL. This property can only be set when using SSL on - self-hosted IR. The default value is the cacerts.pem file installed with the IR. - :type trusted_cert_path: any - :param use_system_trust_store: Specifies whether to use a CA certificate from the system trust - store or from a specified PEM file. The default value is false. - :type use_system_trust_store: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'project': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'project': {'key': 'typeProperties.project', 'type': 'object'}, - 'additional_projects': {'key': 'typeProperties.additionalProjects', 'type': 'object'}, - 'request_google_drive_scope': {'key': 'typeProperties.requestGoogleDriveScope', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'refresh_token': {'key': 'typeProperties.refreshToken', 'type': 'SecretBase'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'email': {'key': 'typeProperties.email', 'type': 'object'}, - 'key_file_path': {'key': 'typeProperties.keyFilePath', 'type': 'object'}, - 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, - 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(GoogleBigQueryLinkedService, self).__init__(**kwargs) - self.type = 'GoogleBigQuery' # type: str - self.project = kwargs['project'] - self.additional_projects = kwargs.get('additional_projects', None) - self.request_google_drive_scope = kwargs.get('request_google_drive_scope', None) - self.authentication_type = kwargs['authentication_type'] - self.refresh_token = kwargs.get('refresh_token', None) - self.client_id = kwargs.get('client_id', None) - self.client_secret = kwargs.get('client_secret', None) - self.email = kwargs.get('email', None) - self.key_file_path = kwargs.get('key_file_path', None) - self.trusted_cert_path = kwargs.get('trusted_cert_path', None) - self.use_system_trust_store = kwargs.get('use_system_trust_store', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class GoogleBigQueryObjectDataset(Dataset): - """Google BigQuery service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using database + table - properties instead. - :type table_name: any - :param table: The table name of the Google BigQuery. Type: string (or Expression with - resultType string). - :type table: any - :param dataset: The database name of the Google BigQuery. Type: string (or Expression with - resultType string). - :type dataset: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'dataset': {'key': 'typeProperties.dataset', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(GoogleBigQueryObjectDataset, self).__init__(**kwargs) - self.type = 'GoogleBigQueryObject' # type: str - self.table_name = kwargs.get('table_name', None) - self.table = kwargs.get('table', None) - self.dataset = kwargs.get('dataset', None) - - -class GoogleBigQuerySource(TabularSource): - """A copy activity Google BigQuery service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(GoogleBigQuerySource, self).__init__(**kwargs) - self.type = 'GoogleBigQuerySource' # type: str - self.query = kwargs.get('query', None) - - -class GoogleCloudStorageLinkedService(LinkedService): - """Linked service for Google Cloud Storage. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param access_key_id: The access key identifier of the Google Cloud Storage Identity and Access - Management (IAM) user. Type: string (or Expression with resultType string). - :type access_key_id: any - :param secret_access_key: The secret access key of the Google Cloud Storage Identity and Access - Management (IAM) user. - :type secret_access_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param service_url: This value specifies the endpoint to access with the Google Cloud Storage - Connector. This is an optional property; change it only if you want to try a different service - endpoint or want to switch between https and http. Type: string (or Expression with resultType - string). - :type service_url: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'access_key_id': {'key': 'typeProperties.accessKeyId', 'type': 'object'}, - 'secret_access_key': {'key': 'typeProperties.secretAccessKey', 'type': 'SecretBase'}, - 'service_url': {'key': 'typeProperties.serviceUrl', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(GoogleCloudStorageLinkedService, self).__init__(**kwargs) - self.type = 'GoogleCloudStorage' # type: str - self.access_key_id = kwargs.get('access_key_id', None) - self.secret_access_key = kwargs.get('secret_access_key', None) - self.service_url = kwargs.get('service_url', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class GoogleCloudStorageLocation(DatasetLocation): - """The location of Google Cloud Storage dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - :param bucket_name: Specify the bucketName of Google Cloud Storage. Type: string (or Expression - with resultType string). - :type bucket_name: any - :param version: Specify the version of Google Cloud Storage. Type: string (or Expression with - resultType string). - :type version: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - 'bucket_name': {'key': 'bucketName', 'type': 'object'}, - 'version': {'key': 'version', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(GoogleCloudStorageLocation, self).__init__(**kwargs) - self.type = 'GoogleCloudStorageLocation' # type: str - self.bucket_name = kwargs.get('bucket_name', None) - self.version = kwargs.get('version', None) - - -class GoogleCloudStorageReadSettings(StoreReadSettings): - """Google Cloud Storage read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: Google Cloud Storage wildcardFolderPath. Type: string (or - Expression with resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: Google Cloud Storage wildcardFileName. Type: string (or Expression - with resultType string). - :type wildcard_file_name: any - :param prefix: The prefix filter for the Google Cloud Storage object name. Type: string (or - Expression with resultType string). - :type prefix: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'prefix': {'key': 'prefix', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - 'modified_datetime_start': {'key': 'modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'modifiedDatetimeEnd', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(GoogleCloudStorageReadSettings, self).__init__(**kwargs) - self.type = 'GoogleCloudStorageReadSettings' # type: str - self.recursive = kwargs.get('recursive', None) - self.wildcard_folder_path = kwargs.get('wildcard_folder_path', None) - self.wildcard_file_name = kwargs.get('wildcard_file_name', None) - self.prefix = kwargs.get('prefix', None) - self.file_list_path = kwargs.get('file_list_path', None) - self.enable_partition_discovery = kwargs.get('enable_partition_discovery', None) - self.partition_root_path = kwargs.get('partition_root_path', None) - self.delete_files_after_completion = kwargs.get('delete_files_after_completion', None) - self.modified_datetime_start = kwargs.get('modified_datetime_start', None) - self.modified_datetime_end = kwargs.get('modified_datetime_end', None) - - -class GreenplumLinkedService(LinkedService): - """Greenplum Database linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: An ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(GreenplumLinkedService, self).__init__(**kwargs) - self.type = 'Greenplum' # type: str - self.connection_string = kwargs.get('connection_string', None) - self.pwd = kwargs.get('pwd', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class GreenplumSource(TabularSource): - """A copy activity Greenplum Database source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(GreenplumSource, self).__init__(**kwargs) - self.type = 'GreenplumSource' # type: str - self.query = kwargs.get('query', None) - - -class GreenplumTableDataset(Dataset): - """Greenplum Database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The table name of Greenplum. Type: string (or Expression with resultType string). - :type table: any - :param schema_type_properties_schema: The schema name of Greenplum. Type: string (or Expression - with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(GreenplumTableDataset, self).__init__(**kwargs) - self.type = 'GreenplumTable' # type: str - self.table_name = kwargs.get('table_name', None) - self.table = kwargs.get('table', None) - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - - -class HBaseLinkedService(LinkedService): - """HBase server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The IP address or host name of the HBase server. (i.e. 192.168.222.160). - :type host: any - :param port: The TCP port that the HBase instance uses to listen for client connections. The - default value is 9090. - :type port: any - :param http_path: The partial URL corresponding to the HBase server. (i.e. - /gateway/sandbox/hbase/version). - :type http_path: any - :param authentication_type: Required. The authentication mechanism to use to connect to the - HBase server. Possible values include: "Anonymous", "Basic". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.HBaseAuthenticationType - :param username: The user name used to connect to the HBase instance. - :type username: any - :param password: The password corresponding to the user name. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The - default value is false. - :type enable_ssl: any - :param trusted_cert_path: The full path of the .pem file containing trusted CA certificates for - verifying the server when connecting over SSL. This property can only be set when using SSL on - self-hosted IR. The default value is the cacerts.pem file installed with the IR. - :type trusted_cert_path: any - :param allow_host_name_cn_mismatch: Specifies whether to require a CA-issued SSL certificate - name to match the host name of the server when connecting over SSL. The default value is false. - :type allow_host_name_cn_mismatch: any - :param allow_self_signed_server_cert: Specifies whether to allow self-signed certificates from - the server. The default value is false. - :type allow_self_signed_server_cert: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'http_path': {'key': 'typeProperties.httpPath', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, - 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, - 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, - 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(HBaseLinkedService, self).__init__(**kwargs) - self.type = 'HBase' # type: str - self.host = kwargs['host'] - self.port = kwargs.get('port', None) - self.http_path = kwargs.get('http_path', None) - self.authentication_type = kwargs['authentication_type'] - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.enable_ssl = kwargs.get('enable_ssl', None) - self.trusted_cert_path = kwargs.get('trusted_cert_path', None) - self.allow_host_name_cn_mismatch = kwargs.get('allow_host_name_cn_mismatch', None) - self.allow_self_signed_server_cert = kwargs.get('allow_self_signed_server_cert', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class HBaseObjectDataset(Dataset): - """HBase server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(HBaseObjectDataset, self).__init__(**kwargs) - self.type = 'HBaseObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class HBaseSource(TabularSource): - """A copy activity HBase server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(HBaseSource, self).__init__(**kwargs) - self.type = 'HBaseSource' # type: str - self.query = kwargs.get('query', None) - - -class HdfsLinkedService(LinkedService): - """Hadoop Distributed File System (HDFS) linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param url: Required. The URL of the HDFS service endpoint, e.g. - http://myhostname:50070/webhdfs/v1 . Type: string (or Expression with resultType string). - :type url: any - :param authentication_type: Type of authentication used to connect to the HDFS. Possible values - are: Anonymous and Windows. Type: string (or Expression with resultType string). - :type authentication_type: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - :param user_name: User name for Windows authentication. Type: string (or Expression with - resultType string). - :type user_name: any - :param password: Password for Windows authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - } - - def __init__( - self, - **kwargs - ): - super(HdfsLinkedService, self).__init__(**kwargs) - self.type = 'Hdfs' # type: str - self.url = kwargs['url'] - self.authentication_type = kwargs.get('authentication_type', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - self.user_name = kwargs.get('user_name', None) - self.password = kwargs.get('password', None) - - -class HdfsLocation(DatasetLocation): - """The location of HDFS. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(HdfsLocation, self).__init__(**kwargs) - self.type = 'HdfsLocation' # type: str - - -class HdfsReadSettings(StoreReadSettings): - """HDFS read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: HDFS wildcardFolderPath. Type: string (or Expression with - resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: HDFS wildcardFileName. Type: string (or Expression with resultType - string). - :type wildcard_file_name: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - :param distcp_settings: Specifies Distcp-related settings. - :type distcp_settings: ~azure.synapse.artifacts.v2020_12_01.models.DistcpSettings - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'modified_datetime_start': {'key': 'modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'modifiedDatetimeEnd', 'type': 'object'}, - 'distcp_settings': {'key': 'distcpSettings', 'type': 'DistcpSettings'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(HdfsReadSettings, self).__init__(**kwargs) - self.type = 'HdfsReadSettings' # type: str - self.recursive = kwargs.get('recursive', None) - self.wildcard_folder_path = kwargs.get('wildcard_folder_path', None) - self.wildcard_file_name = kwargs.get('wildcard_file_name', None) - self.file_list_path = kwargs.get('file_list_path', None) - self.enable_partition_discovery = kwargs.get('enable_partition_discovery', None) - self.partition_root_path = kwargs.get('partition_root_path', None) - self.modified_datetime_start = kwargs.get('modified_datetime_start', None) - self.modified_datetime_end = kwargs.get('modified_datetime_end', None) - self.distcp_settings = kwargs.get('distcp_settings', None) - self.delete_files_after_completion = kwargs.get('delete_files_after_completion', None) - - -class HdfsSource(CopySource): - """A copy activity HDFS source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param distcp_settings: Specifies Distcp-related settings. - :type distcp_settings: ~azure.synapse.artifacts.v2020_12_01.models.DistcpSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'distcp_settings': {'key': 'distcpSettings', 'type': 'DistcpSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(HdfsSource, self).__init__(**kwargs) - self.type = 'HdfsSource' # type: str - self.recursive = kwargs.get('recursive', None) - self.distcp_settings = kwargs.get('distcp_settings', None) - - -class HDInsightHiveActivity(ExecutionActivity): - """HDInsight Hive activity type. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param storage_linked_services: Storage linked service references. - :type storage_linked_services: - list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference] - :param arguments: User specified arguments to HDInsightActivity. - :type arguments: list[any] - :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2020_12_01.models.HDInsightActivityDebugInfoOption - :param script_path: Script path. Type: string (or Expression with resultType string). - :type script_path: any - :param script_linked_service: Script linked service reference. - :type script_linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param defines: Allows user to specify defines for Hive job request. - :type defines: dict[str, any] - :param variables: User specified arguments under hivevar namespace. - :type variables: list[any] - :param query_timeout: Query timeout value (in minutes). Effective when the HDInsight cluster - is with ESP (Enterprise Security Package). - :type query_timeout: int - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'storage_linked_services': {'key': 'typeProperties.storageLinkedServices', 'type': '[LinkedServiceReference]'}, - 'arguments': {'key': 'typeProperties.arguments', 'type': '[object]'}, - 'get_debug_info': {'key': 'typeProperties.getDebugInfo', 'type': 'str'}, - 'script_path': {'key': 'typeProperties.scriptPath', 'type': 'object'}, - 'script_linked_service': {'key': 'typeProperties.scriptLinkedService', 'type': 'LinkedServiceReference'}, - 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, - 'variables': {'key': 'typeProperties.variables', 'type': '[object]'}, - 'query_timeout': {'key': 'typeProperties.queryTimeout', 'type': 'int'}, - } - - def __init__( - self, - **kwargs - ): - super(HDInsightHiveActivity, self).__init__(**kwargs) - self.type = 'HDInsightHive' # type: str - self.storage_linked_services = kwargs.get('storage_linked_services', None) - self.arguments = kwargs.get('arguments', None) - self.get_debug_info = kwargs.get('get_debug_info', None) - self.script_path = kwargs.get('script_path', None) - self.script_linked_service = kwargs.get('script_linked_service', None) - self.defines = kwargs.get('defines', None) - self.variables = kwargs.get('variables', None) - self.query_timeout = kwargs.get('query_timeout', None) - - -class HDInsightLinkedService(LinkedService): - """HDInsight linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param cluster_uri: Required. HDInsight cluster URI. Type: string (or Expression with - resultType string). - :type cluster_uri: any - :param user_name: HDInsight cluster user name. Type: string (or Expression with resultType - string). - :type user_name: any - :param password: HDInsight cluster password. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param linked_service_name: The Azure Storage linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param hcatalog_linked_service_name: A reference to the Azure SQL linked service that points to - the HCatalog database. - :type hcatalog_linked_service_name: - ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - :param is_esp_enabled: Specify if the HDInsight is created with ESP (Enterprise Security - Package). Type: Boolean. - :type is_esp_enabled: any - :param file_system: Specify the FileSystem if the main storage for the HDInsight is ADLS Gen2. - Type: string (or Expression with resultType string). - :type file_system: any - """ - - _validation = { - 'type': {'required': True}, - 'cluster_uri': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'cluster_uri': {'key': 'typeProperties.clusterUri', 'type': 'object'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'linked_service_name': {'key': 'typeProperties.linkedServiceName', 'type': 'LinkedServiceReference'}, - 'hcatalog_linked_service_name': {'key': 'typeProperties.hcatalogLinkedServiceName', 'type': 'LinkedServiceReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - 'is_esp_enabled': {'key': 'typeProperties.isEspEnabled', 'type': 'object'}, - 'file_system': {'key': 'typeProperties.fileSystem', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(HDInsightLinkedService, self).__init__(**kwargs) - self.type = 'HDInsight' # type: str - self.cluster_uri = kwargs['cluster_uri'] - self.user_name = kwargs.get('user_name', None) - self.password = kwargs.get('password', None) - self.linked_service_name = kwargs.get('linked_service_name', None) - self.hcatalog_linked_service_name = kwargs.get('hcatalog_linked_service_name', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - self.is_esp_enabled = kwargs.get('is_esp_enabled', None) - self.file_system = kwargs.get('file_system', None) - - -class HDInsightMapReduceActivity(ExecutionActivity): - """HDInsight MapReduce activity type. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param storage_linked_services: Storage linked service references. - :type storage_linked_services: - list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference] - :param arguments: User specified arguments to HDInsightActivity. - :type arguments: list[any] - :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2020_12_01.models.HDInsightActivityDebugInfoOption - :param class_name: Required. Class name. Type: string (or Expression with resultType string). - :type class_name: any - :param jar_file_path: Required. Jar path. Type: string (or Expression with resultType string). - :type jar_file_path: any - :param jar_linked_service: Jar linked service reference. - :type jar_linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param jar_libs: Jar libs. - :type jar_libs: list[any] - :param defines: Allows user to specify defines for the MapReduce job request. - :type defines: dict[str, any] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'class_name': {'required': True}, - 'jar_file_path': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'storage_linked_services': {'key': 'typeProperties.storageLinkedServices', 'type': '[LinkedServiceReference]'}, - 'arguments': {'key': 'typeProperties.arguments', 'type': '[object]'}, - 'get_debug_info': {'key': 'typeProperties.getDebugInfo', 'type': 'str'}, - 'class_name': {'key': 'typeProperties.className', 'type': 'object'}, - 'jar_file_path': {'key': 'typeProperties.jarFilePath', 'type': 'object'}, - 'jar_linked_service': {'key': 'typeProperties.jarLinkedService', 'type': 'LinkedServiceReference'}, - 'jar_libs': {'key': 'typeProperties.jarLibs', 'type': '[object]'}, - 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, - } - - def __init__( - self, - **kwargs - ): - super(HDInsightMapReduceActivity, self).__init__(**kwargs) - self.type = 'HDInsightMapReduce' # type: str - self.storage_linked_services = kwargs.get('storage_linked_services', None) - self.arguments = kwargs.get('arguments', None) - self.get_debug_info = kwargs.get('get_debug_info', None) - self.class_name = kwargs['class_name'] - self.jar_file_path = kwargs['jar_file_path'] - self.jar_linked_service = kwargs.get('jar_linked_service', None) - self.jar_libs = kwargs.get('jar_libs', None) - self.defines = kwargs.get('defines', None) - - -class HDInsightOnDemandLinkedService(LinkedService): - """HDInsight ondemand linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param cluster_size: Required. Number of worker/data nodes in the cluster. Suggestion value: 4. - Type: string (or Expression with resultType string). - :type cluster_size: any - :param time_to_live: Required. The allowed idle time for the on-demand HDInsight cluster. - Specifies how long the on-demand HDInsight cluster stays alive after completion of an activity - run if there are no other active jobs in the cluster. The minimum value is 5 mins. Type: string - (or Expression with resultType string). - :type time_to_live: any - :param version: Required. Version of the HDInsight cluster.  Type: string (or Expression with - resultType string). - :type version: any - :param linked_service_name: Required. Azure Storage linked service to be used by the on-demand - cluster for storing and processing data. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param host_subscription_id: Required. The customer’s subscription to host the cluster. Type: - string (or Expression with resultType string). - :type host_subscription_id: any - :param service_principal_id: The service principal id for the hostSubscriptionId. Type: string - (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: The key for the service principal id. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: Required. The Tenant id/name to which the service principal belongs. Type: - string (or Expression with resultType string). - :type tenant: any - :param cluster_resource_group: Required. The resource group where the cluster belongs. Type: - string (or Expression with resultType string). - :type cluster_resource_group: any - :param cluster_name_prefix: The prefix of cluster name, postfix will be distinct with - timestamp. Type: string (or Expression with resultType string). - :type cluster_name_prefix: any - :param cluster_user_name: The username to access the cluster. Type: string (or Expression with - resultType string). - :type cluster_user_name: any - :param cluster_password: The password to access the cluster. - :type cluster_password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param cluster_ssh_user_name: The username to SSH remotely connect to cluster’s node (for - Linux). Type: string (or Expression with resultType string). - :type cluster_ssh_user_name: any - :param cluster_ssh_password: The password to SSH remotely connect cluster’s node (for Linux). - :type cluster_ssh_password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param additional_linked_service_names: Specifies additional storage accounts for the HDInsight - linked service so that the Data Factory service can register them on your behalf. - :type additional_linked_service_names: - list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference] - :param hcatalog_linked_service_name: The name of Azure SQL linked service that point to the - HCatalog database. The on-demand HDInsight cluster is created by using the Azure SQL database - as the metastore. - :type hcatalog_linked_service_name: - ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param cluster_type: The cluster type. Type: string (or Expression with resultType string). - :type cluster_type: any - :param spark_version: The version of spark if the cluster type is 'spark'. Type: string (or - Expression with resultType string). - :type spark_version: any - :param core_configuration: Specifies the core configuration parameters (as in core-site.xml) - for the HDInsight cluster to be created. - :type core_configuration: any - :param h_base_configuration: Specifies the HBase configuration parameters (hbase-site.xml) for - the HDInsight cluster. - :type h_base_configuration: any - :param hdfs_configuration: Specifies the HDFS configuration parameters (hdfs-site.xml) for the - HDInsight cluster. - :type hdfs_configuration: any - :param hive_configuration: Specifies the hive configuration parameters (hive-site.xml) for the - HDInsight cluster. - :type hive_configuration: any - :param map_reduce_configuration: Specifies the MapReduce configuration parameters - (mapred-site.xml) for the HDInsight cluster. - :type map_reduce_configuration: any - :param oozie_configuration: Specifies the Oozie configuration parameters (oozie-site.xml) for - the HDInsight cluster. - :type oozie_configuration: any - :param storm_configuration: Specifies the Storm configuration parameters (storm-site.xml) for - the HDInsight cluster. - :type storm_configuration: any - :param yarn_configuration: Specifies the Yarn configuration parameters (yarn-site.xml) for the - HDInsight cluster. - :type yarn_configuration: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - :param head_node_size: Specifies the size of the head node for the HDInsight cluster. - :type head_node_size: any - :param data_node_size: Specifies the size of the data node for the HDInsight cluster. - :type data_node_size: any - :param zookeeper_node_size: Specifies the size of the Zoo Keeper node for the HDInsight - cluster. - :type zookeeper_node_size: any - :param script_actions: Custom script actions to run on HDI ondemand cluster once it's up. - Please refer to - https://docs.microsoft.com/en-us/azure/hdinsight/hdinsight-hadoop-customize-cluster-linux?toc=%2Fen-us%2Fazure%2Fhdinsight%2Fr-server%2FTOC.json&bc=%2Fen-us%2Fazure%2Fbread%2Ftoc.json#understanding-script-actions. - :type script_actions: list[~azure.synapse.artifacts.v2020_12_01.models.ScriptAction] - :param virtual_network_id: The ARM resource ID for the vNet to which the cluster should be - joined after creation. Type: string (or Expression with resultType string). - :type virtual_network_id: any - :param subnet_name: The ARM resource ID for the subnet in the vNet. If virtualNetworkId was - specified, then this property is required. Type: string (or Expression with resultType string). - :type subnet_name: any - """ - - _validation = { - 'type': {'required': True}, - 'cluster_size': {'required': True}, - 'time_to_live': {'required': True}, - 'version': {'required': True}, - 'linked_service_name': {'required': True}, - 'host_subscription_id': {'required': True}, - 'tenant': {'required': True}, - 'cluster_resource_group': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'cluster_size': {'key': 'typeProperties.clusterSize', 'type': 'object'}, - 'time_to_live': {'key': 'typeProperties.timeToLive', 'type': 'object'}, - 'version': {'key': 'typeProperties.version', 'type': 'object'}, - 'linked_service_name': {'key': 'typeProperties.linkedServiceName', 'type': 'LinkedServiceReference'}, - 'host_subscription_id': {'key': 'typeProperties.hostSubscriptionId', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'cluster_resource_group': {'key': 'typeProperties.clusterResourceGroup', 'type': 'object'}, - 'cluster_name_prefix': {'key': 'typeProperties.clusterNamePrefix', 'type': 'object'}, - 'cluster_user_name': {'key': 'typeProperties.clusterUserName', 'type': 'object'}, - 'cluster_password': {'key': 'typeProperties.clusterPassword', 'type': 'SecretBase'}, - 'cluster_ssh_user_name': {'key': 'typeProperties.clusterSshUserName', 'type': 'object'}, - 'cluster_ssh_password': {'key': 'typeProperties.clusterSshPassword', 'type': 'SecretBase'}, - 'additional_linked_service_names': {'key': 'typeProperties.additionalLinkedServiceNames', 'type': '[LinkedServiceReference]'}, - 'hcatalog_linked_service_name': {'key': 'typeProperties.hcatalogLinkedServiceName', 'type': 'LinkedServiceReference'}, - 'cluster_type': {'key': 'typeProperties.clusterType', 'type': 'object'}, - 'spark_version': {'key': 'typeProperties.sparkVersion', 'type': 'object'}, - 'core_configuration': {'key': 'typeProperties.coreConfiguration', 'type': 'object'}, - 'h_base_configuration': {'key': 'typeProperties.hBaseConfiguration', 'type': 'object'}, - 'hdfs_configuration': {'key': 'typeProperties.hdfsConfiguration', 'type': 'object'}, - 'hive_configuration': {'key': 'typeProperties.hiveConfiguration', 'type': 'object'}, - 'map_reduce_configuration': {'key': 'typeProperties.mapReduceConfiguration', 'type': 'object'}, - 'oozie_configuration': {'key': 'typeProperties.oozieConfiguration', 'type': 'object'}, - 'storm_configuration': {'key': 'typeProperties.stormConfiguration', 'type': 'object'}, - 'yarn_configuration': {'key': 'typeProperties.yarnConfiguration', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - 'head_node_size': {'key': 'typeProperties.headNodeSize', 'type': 'object'}, - 'data_node_size': {'key': 'typeProperties.dataNodeSize', 'type': 'object'}, - 'zookeeper_node_size': {'key': 'typeProperties.zookeeperNodeSize', 'type': 'object'}, - 'script_actions': {'key': 'typeProperties.scriptActions', 'type': '[ScriptAction]'}, - 'virtual_network_id': {'key': 'typeProperties.virtualNetworkId', 'type': 'object'}, - 'subnet_name': {'key': 'typeProperties.subnetName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(HDInsightOnDemandLinkedService, self).__init__(**kwargs) - self.type = 'HDInsightOnDemand' # type: str - self.cluster_size = kwargs['cluster_size'] - self.time_to_live = kwargs['time_to_live'] - self.version = kwargs['version'] - self.linked_service_name = kwargs['linked_service_name'] - self.host_subscription_id = kwargs['host_subscription_id'] - self.service_principal_id = kwargs.get('service_principal_id', None) - self.service_principal_key = kwargs.get('service_principal_key', None) - self.tenant = kwargs['tenant'] - self.cluster_resource_group = kwargs['cluster_resource_group'] - self.cluster_name_prefix = kwargs.get('cluster_name_prefix', None) - self.cluster_user_name = kwargs.get('cluster_user_name', None) - self.cluster_password = kwargs.get('cluster_password', None) - self.cluster_ssh_user_name = kwargs.get('cluster_ssh_user_name', None) - self.cluster_ssh_password = kwargs.get('cluster_ssh_password', None) - self.additional_linked_service_names = kwargs.get('additional_linked_service_names', None) - self.hcatalog_linked_service_name = kwargs.get('hcatalog_linked_service_name', None) - self.cluster_type = kwargs.get('cluster_type', None) - self.spark_version = kwargs.get('spark_version', None) - self.core_configuration = kwargs.get('core_configuration', None) - self.h_base_configuration = kwargs.get('h_base_configuration', None) - self.hdfs_configuration = kwargs.get('hdfs_configuration', None) - self.hive_configuration = kwargs.get('hive_configuration', None) - self.map_reduce_configuration = kwargs.get('map_reduce_configuration', None) - self.oozie_configuration = kwargs.get('oozie_configuration', None) - self.storm_configuration = kwargs.get('storm_configuration', None) - self.yarn_configuration = kwargs.get('yarn_configuration', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - self.head_node_size = kwargs.get('head_node_size', None) - self.data_node_size = kwargs.get('data_node_size', None) - self.zookeeper_node_size = kwargs.get('zookeeper_node_size', None) - self.script_actions = kwargs.get('script_actions', None) - self.virtual_network_id = kwargs.get('virtual_network_id', None) - self.subnet_name = kwargs.get('subnet_name', None) - - -class HDInsightPigActivity(ExecutionActivity): - """HDInsight Pig activity type. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param storage_linked_services: Storage linked service references. - :type storage_linked_services: - list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference] - :param arguments: User specified arguments to HDInsightActivity. Type: array (or Expression - with resultType array). - :type arguments: any - :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2020_12_01.models.HDInsightActivityDebugInfoOption - :param script_path: Script path. Type: string (or Expression with resultType string). - :type script_path: any - :param script_linked_service: Script linked service reference. - :type script_linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param defines: Allows user to specify defines for Pig job request. - :type defines: dict[str, any] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'storage_linked_services': {'key': 'typeProperties.storageLinkedServices', 'type': '[LinkedServiceReference]'}, - 'arguments': {'key': 'typeProperties.arguments', 'type': 'object'}, - 'get_debug_info': {'key': 'typeProperties.getDebugInfo', 'type': 'str'}, - 'script_path': {'key': 'typeProperties.scriptPath', 'type': 'object'}, - 'script_linked_service': {'key': 'typeProperties.scriptLinkedService', 'type': 'LinkedServiceReference'}, - 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, - } - - def __init__( - self, - **kwargs - ): - super(HDInsightPigActivity, self).__init__(**kwargs) - self.type = 'HDInsightPig' # type: str - self.storage_linked_services = kwargs.get('storage_linked_services', None) - self.arguments = kwargs.get('arguments', None) - self.get_debug_info = kwargs.get('get_debug_info', None) - self.script_path = kwargs.get('script_path', None) - self.script_linked_service = kwargs.get('script_linked_service', None) - self.defines = kwargs.get('defines', None) - - -class HDInsightSparkActivity(ExecutionActivity): - """HDInsight Spark activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param root_path: Required. The root path in 'sparkJobLinkedService' for all the job’s files. - Type: string (or Expression with resultType string). - :type root_path: any - :param entry_file_path: Required. The relative path to the root folder of the code/package to - be executed. Type: string (or Expression with resultType string). - :type entry_file_path: any - :param arguments: The user-specified arguments to HDInsightSparkActivity. - :type arguments: list[any] - :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2020_12_01.models.HDInsightActivityDebugInfoOption - :param spark_job_linked_service: The storage linked service for uploading the entry file and - dependencies, and for receiving logs. - :type spark_job_linked_service: - ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param class_name: The application's Java/Spark main class. - :type class_name: str - :param proxy_user: The user to impersonate that will execute the job. Type: string (or - Expression with resultType string). - :type proxy_user: any - :param spark_config: Spark configuration property. - :type spark_config: dict[str, any] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'root_path': {'required': True}, - 'entry_file_path': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'root_path': {'key': 'typeProperties.rootPath', 'type': 'object'}, - 'entry_file_path': {'key': 'typeProperties.entryFilePath', 'type': 'object'}, - 'arguments': {'key': 'typeProperties.arguments', 'type': '[object]'}, - 'get_debug_info': {'key': 'typeProperties.getDebugInfo', 'type': 'str'}, - 'spark_job_linked_service': {'key': 'typeProperties.sparkJobLinkedService', 'type': 'LinkedServiceReference'}, - 'class_name': {'key': 'typeProperties.className', 'type': 'str'}, - 'proxy_user': {'key': 'typeProperties.proxyUser', 'type': 'object'}, - 'spark_config': {'key': 'typeProperties.sparkConfig', 'type': '{object}'}, - } - - def __init__( - self, - **kwargs - ): - super(HDInsightSparkActivity, self).__init__(**kwargs) - self.type = 'HDInsightSpark' # type: str - self.root_path = kwargs['root_path'] - self.entry_file_path = kwargs['entry_file_path'] - self.arguments = kwargs.get('arguments', None) - self.get_debug_info = kwargs.get('get_debug_info', None) - self.spark_job_linked_service = kwargs.get('spark_job_linked_service', None) - self.class_name = kwargs.get('class_name', None) - self.proxy_user = kwargs.get('proxy_user', None) - self.spark_config = kwargs.get('spark_config', None) - - -class HDInsightStreamingActivity(ExecutionActivity): - """HDInsight streaming activity type. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param storage_linked_services: Storage linked service references. - :type storage_linked_services: - list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference] - :param arguments: User specified arguments to HDInsightActivity. - :type arguments: list[any] - :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2020_12_01.models.HDInsightActivityDebugInfoOption - :param mapper: Required. Mapper executable name. Type: string (or Expression with resultType - string). - :type mapper: any - :param reducer: Required. Reducer executable name. Type: string (or Expression with resultType - string). - :type reducer: any - :param input: Required. Input blob path. Type: string (or Expression with resultType string). - :type input: any - :param output: Required. Output blob path. Type: string (or Expression with resultType string). - :type output: any - :param file_paths: Required. Paths to streaming job files. Can be directories. - :type file_paths: list[any] - :param file_linked_service: Linked service reference where the files are located. - :type file_linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param combiner: Combiner executable name. Type: string (or Expression with resultType string). - :type combiner: any - :param command_environment: Command line environment values. - :type command_environment: list[any] - :param defines: Allows user to specify defines for streaming job request. - :type defines: dict[str, any] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'mapper': {'required': True}, - 'reducer': {'required': True}, - 'input': {'required': True}, - 'output': {'required': True}, - 'file_paths': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'storage_linked_services': {'key': 'typeProperties.storageLinkedServices', 'type': '[LinkedServiceReference]'}, - 'arguments': {'key': 'typeProperties.arguments', 'type': '[object]'}, - 'get_debug_info': {'key': 'typeProperties.getDebugInfo', 'type': 'str'}, - 'mapper': {'key': 'typeProperties.mapper', 'type': 'object'}, - 'reducer': {'key': 'typeProperties.reducer', 'type': 'object'}, - 'input': {'key': 'typeProperties.input', 'type': 'object'}, - 'output': {'key': 'typeProperties.output', 'type': 'object'}, - 'file_paths': {'key': 'typeProperties.filePaths', 'type': '[object]'}, - 'file_linked_service': {'key': 'typeProperties.fileLinkedService', 'type': 'LinkedServiceReference'}, - 'combiner': {'key': 'typeProperties.combiner', 'type': 'object'}, - 'command_environment': {'key': 'typeProperties.commandEnvironment', 'type': '[object]'}, - 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, - } - - def __init__( - self, - **kwargs - ): - super(HDInsightStreamingActivity, self).__init__(**kwargs) - self.type = 'HDInsightStreaming' # type: str - self.storage_linked_services = kwargs.get('storage_linked_services', None) - self.arguments = kwargs.get('arguments', None) - self.get_debug_info = kwargs.get('get_debug_info', None) - self.mapper = kwargs['mapper'] - self.reducer = kwargs['reducer'] - self.input = kwargs['input'] - self.output = kwargs['output'] - self.file_paths = kwargs['file_paths'] - self.file_linked_service = kwargs.get('file_linked_service', None) - self.combiner = kwargs.get('combiner', None) - self.command_environment = kwargs.get('command_environment', None) - self.defines = kwargs.get('defines', None) - - -class HiveLinkedService(LinkedService): - """Hive Server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. IP address or host name of the Hive server, separated by ';' for - multiple hosts (only when serviceDiscoveryMode is enable). - :type host: any - :param port: The TCP port that the Hive server uses to listen for client connections. - :type port: any - :param server_type: The type of Hive server. Possible values include: "HiveServer1", - "HiveServer2", "HiveThriftServer". - :type server_type: str or ~azure.synapse.artifacts.v2020_12_01.models.HiveServerType - :param thrift_transport_protocol: The transport protocol to use in the Thrift layer. Possible - values include: "Binary", "SASL", "HTTP ". - :type thrift_transport_protocol: str or - ~azure.synapse.artifacts.v2020_12_01.models.HiveThriftTransportProtocol - :param authentication_type: Required. The authentication method used to access the Hive server. - Possible values include: "Anonymous", "Username", "UsernameAndPassword", - "WindowsAzureHDInsightService". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.HiveAuthenticationType - :param service_discovery_mode: true to indicate using the ZooKeeper service, false not. - :type service_discovery_mode: any - :param zoo_keeper_name_space: The namespace on ZooKeeper under which Hive Server 2 nodes are - added. - :type zoo_keeper_name_space: any - :param use_native_query: Specifies whether the driver uses native HiveQL queries,or converts - them into an equivalent form in HiveQL. - :type use_native_query: any - :param username: The user name that you use to access Hive Server. - :type username: any - :param password: The password corresponding to the user name that you provided in the Username - field. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param http_path: The partial URL corresponding to the Hive server. - :type http_path: any - :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The - default value is false. - :type enable_ssl: any - :param trusted_cert_path: The full path of the .pem file containing trusted CA certificates for - verifying the server when connecting over SSL. This property can only be set when using SSL on - self-hosted IR. The default value is the cacerts.pem file installed with the IR. - :type trusted_cert_path: any - :param use_system_trust_store: Specifies whether to use a CA certificate from the system trust - store or from a specified PEM file. The default value is false. - :type use_system_trust_store: any - :param allow_host_name_cn_mismatch: Specifies whether to require a CA-issued SSL certificate - name to match the host name of the server when connecting over SSL. The default value is false. - :type allow_host_name_cn_mismatch: any - :param allow_self_signed_server_cert: Specifies whether to allow self-signed certificates from - the server. The default value is false. - :type allow_self_signed_server_cert: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'server_type': {'key': 'typeProperties.serverType', 'type': 'str'}, - 'thrift_transport_protocol': {'key': 'typeProperties.thriftTransportProtocol', 'type': 'str'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'service_discovery_mode': {'key': 'typeProperties.serviceDiscoveryMode', 'type': 'object'}, - 'zoo_keeper_name_space': {'key': 'typeProperties.zooKeeperNameSpace', 'type': 'object'}, - 'use_native_query': {'key': 'typeProperties.useNativeQuery', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'http_path': {'key': 'typeProperties.httpPath', 'type': 'object'}, - 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, - 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, - 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, - 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, - 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(HiveLinkedService, self).__init__(**kwargs) - self.type = 'Hive' # type: str - self.host = kwargs['host'] - self.port = kwargs.get('port', None) - self.server_type = kwargs.get('server_type', None) - self.thrift_transport_protocol = kwargs.get('thrift_transport_protocol', None) - self.authentication_type = kwargs['authentication_type'] - self.service_discovery_mode = kwargs.get('service_discovery_mode', None) - self.zoo_keeper_name_space = kwargs.get('zoo_keeper_name_space', None) - self.use_native_query = kwargs.get('use_native_query', None) - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.http_path = kwargs.get('http_path', None) - self.enable_ssl = kwargs.get('enable_ssl', None) - self.trusted_cert_path = kwargs.get('trusted_cert_path', None) - self.use_system_trust_store = kwargs.get('use_system_trust_store', None) - self.allow_host_name_cn_mismatch = kwargs.get('allow_host_name_cn_mismatch', None) - self.allow_self_signed_server_cert = kwargs.get('allow_self_signed_server_cert', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class HiveObjectDataset(Dataset): - """Hive Server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The table name of the Hive. Type: string (or Expression with resultType string). - :type table: any - :param schema_type_properties_schema: The schema name of the Hive. Type: string (or Expression - with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(HiveObjectDataset, self).__init__(**kwargs) - self.type = 'HiveObject' # type: str - self.table_name = kwargs.get('table_name', None) - self.table = kwargs.get('table', None) - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - - -class HiveSource(TabularSource): - """A copy activity Hive Server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(HiveSource, self).__init__(**kwargs) - self.type = 'HiveSource' # type: str - self.query = kwargs.get('query', None) - - -class HttpDataset(Dataset): - """A file in an HTTP web server. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param relative_url: The relative URL based on the URL in the HttpLinkedService refers to an - HTTP file Type: string (or Expression with resultType string). - :type relative_url: any - :param request_method: The HTTP method for the HTTP request. Type: string (or Expression with - resultType string). - :type request_method: any - :param request_body: The body for the HTTP request. Type: string (or Expression with resultType - string). - :type request_body: any - :param additional_headers: The headers for the HTTP Request. e.g. - request-header-name-1:request-header-value-1 - ... - request-header-name-n:request-header-value-n Type: string (or Expression with resultType - string). - :type additional_headers: any - :param format: The format of files. - :type format: ~azure.synapse.artifacts.v2020_12_01.models.DatasetStorageFormat - :param compression: The data compression method used on files. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'relative_url': {'key': 'typeProperties.relativeUrl', 'type': 'object'}, - 'request_method': {'key': 'typeProperties.requestMethod', 'type': 'object'}, - 'request_body': {'key': 'typeProperties.requestBody', 'type': 'object'}, - 'additional_headers': {'key': 'typeProperties.additionalHeaders', 'type': 'object'}, - 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - } - - def __init__( - self, - **kwargs - ): - super(HttpDataset, self).__init__(**kwargs) - self.type = 'HttpFile' # type: str - self.relative_url = kwargs.get('relative_url', None) - self.request_method = kwargs.get('request_method', None) - self.request_body = kwargs.get('request_body', None) - self.additional_headers = kwargs.get('additional_headers', None) - self.format = kwargs.get('format', None) - self.compression = kwargs.get('compression', None) - - -class HttpLinkedService(LinkedService): - """Linked service for an HTTP source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param url: Required. The base URL of the HTTP endpoint, e.g. http://www.microsoft.com. Type: - string (or Expression with resultType string). - :type url: any - :param authentication_type: The authentication type to be used to connect to the HTTP server. - Possible values include: "Basic", "Anonymous", "Digest", "Windows", "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.HttpAuthenticationType - :param user_name: User name for Basic, Digest, or Windows authentication. Type: string (or - Expression with resultType string). - :type user_name: any - :param password: Password for Basic, Digest, Windows, or ClientCertificate with - EmbeddedCertData authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param embedded_cert_data: Base64 encoded certificate data for ClientCertificate - authentication. For on-premises copy with ClientCertificate authentication, either - CertThumbprint or EmbeddedCertData/Password should be specified. Type: string (or Expression - with resultType string). - :type embedded_cert_data: any - :param cert_thumbprint: Thumbprint of certificate for ClientCertificate authentication. Only - valid for on-premises copy. For on-premises copy with ClientCertificate authentication, either - CertThumbprint or EmbeddedCertData/Password should be specified. Type: string (or Expression - with resultType string). - :type cert_thumbprint: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - :param enable_server_certificate_validation: If true, validate the HTTPS server SSL - certificate. Default value is true. Type: boolean (or Expression with resultType boolean). - :type enable_server_certificate_validation: any - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'embedded_cert_data': {'key': 'typeProperties.embeddedCertData', 'type': 'object'}, - 'cert_thumbprint': {'key': 'typeProperties.certThumbprint', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - 'enable_server_certificate_validation': {'key': 'typeProperties.enableServerCertificateValidation', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(HttpLinkedService, self).__init__(**kwargs) - self.type = 'HttpServer' # type: str - self.url = kwargs['url'] - self.authentication_type = kwargs.get('authentication_type', None) - self.user_name = kwargs.get('user_name', None) - self.password = kwargs.get('password', None) - self.embedded_cert_data = kwargs.get('embedded_cert_data', None) - self.cert_thumbprint = kwargs.get('cert_thumbprint', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - self.enable_server_certificate_validation = kwargs.get('enable_server_certificate_validation', None) - - -class HttpReadSettings(StoreReadSettings): - """Sftp read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param request_method: The HTTP method used to call the RESTful API. The default is GET. Type: - string (or Expression with resultType string). - :type request_method: any - :param request_body: The HTTP request body to the RESTful API if requestMethod is POST. Type: - string (or Expression with resultType string). - :type request_body: any - :param additional_headers: The additional HTTP headers in the request to the RESTful API. Type: - string (or Expression with resultType string). - :type additional_headers: any - :param request_timeout: Specifies the timeout for a HTTP client to get HTTP response from HTTP - server. - :type request_timeout: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'request_method': {'key': 'requestMethod', 'type': 'object'}, - 'request_body': {'key': 'requestBody', 'type': 'object'}, - 'additional_headers': {'key': 'additionalHeaders', 'type': 'object'}, - 'request_timeout': {'key': 'requestTimeout', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(HttpReadSettings, self).__init__(**kwargs) - self.type = 'HttpReadSettings' # type: str - self.request_method = kwargs.get('request_method', None) - self.request_body = kwargs.get('request_body', None) - self.additional_headers = kwargs.get('additional_headers', None) - self.request_timeout = kwargs.get('request_timeout', None) - self.enable_partition_discovery = kwargs.get('enable_partition_discovery', None) - self.partition_root_path = kwargs.get('partition_root_path', None) - - -class HttpServerLocation(DatasetLocation): - """The location of http server. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - :param relative_url: Specify the relativeUrl of http server. Type: string (or Expression with - resultType string). - :type relative_url: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - 'relative_url': {'key': 'relativeUrl', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(HttpServerLocation, self).__init__(**kwargs) - self.type = 'HttpServerLocation' # type: str - self.relative_url = kwargs.get('relative_url', None) - - -class HttpSource(CopySource): - """A copy activity source for an HTTP file. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param http_request_timeout: Specifies the timeout for a HTTP client to get HTTP response from - HTTP server. The default value is equivalent to System.Net.HttpWebRequest.Timeout. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(HttpSource, self).__init__(**kwargs) - self.type = 'HttpSource' # type: str - self.http_request_timeout = kwargs.get('http_request_timeout', None) - - -class HubspotLinkedService(LinkedService): - """Hubspot Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param client_id: Required. The client ID associated with your Hubspot application. - :type client_id: any - :param client_secret: The client secret associated with your Hubspot application. - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param access_token: The access token obtained when initially authenticating your OAuth - integration. - :type access_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param refresh_token: The refresh token obtained when initially authenticating your OAuth - integration. - :type refresh_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'client_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, - 'refresh_token': {'key': 'typeProperties.refreshToken', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(HubspotLinkedService, self).__init__(**kwargs) - self.type = 'Hubspot' # type: str - self.client_id = kwargs['client_id'] - self.client_secret = kwargs.get('client_secret', None) - self.access_token = kwargs.get('access_token', None) - self.refresh_token = kwargs.get('refresh_token', None) - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.use_host_verification = kwargs.get('use_host_verification', None) - self.use_peer_verification = kwargs.get('use_peer_verification', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class HubspotObjectDataset(Dataset): - """Hubspot Service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(HubspotObjectDataset, self).__init__(**kwargs) - self.type = 'HubspotObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class HubspotSource(TabularSource): - """A copy activity Hubspot Service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(HubspotSource, self).__init__(**kwargs) - self.type = 'HubspotSource' # type: str - self.query = kwargs.get('query', None) - - -class IfConditionActivity(ControlActivity): - """This activity evaluates a boolean expression and executes either the activities under the ifTrueActivities property or the ifFalseActivities property depending on the result of the expression. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param expression: Required. An expression that would evaluate to Boolean. This is used to - determine the block of activities (ifTrueActivities or ifFalseActivities) that will be - executed. - :type expression: ~azure.synapse.artifacts.v2020_12_01.models.Expression - :param if_true_activities: List of activities to execute if expression is evaluated to true. - This is an optional property and if not provided, the activity will exit without any action. - :type if_true_activities: list[~azure.synapse.artifacts.v2020_12_01.models.Activity] - :param if_false_activities: List of activities to execute if expression is evaluated to false. - This is an optional property and if not provided, the activity will exit without any action. - :type if_false_activities: list[~azure.synapse.artifacts.v2020_12_01.models.Activity] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'expression': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'expression': {'key': 'typeProperties.expression', 'type': 'Expression'}, - 'if_true_activities': {'key': 'typeProperties.ifTrueActivities', 'type': '[Activity]'}, - 'if_false_activities': {'key': 'typeProperties.ifFalseActivities', 'type': '[Activity]'}, - } - - def __init__( - self, - **kwargs - ): - super(IfConditionActivity, self).__init__(**kwargs) - self.type = 'IfCondition' # type: str - self.expression = kwargs['expression'] - self.if_true_activities = kwargs.get('if_true_activities', None) - self.if_false_activities = kwargs.get('if_false_activities', None) - - -class ImpalaLinkedService(LinkedService): - """Impala server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The IP address or host name of the Impala server. (i.e. - 192.168.222.160). - :type host: any - :param port: The TCP port that the Impala server uses to listen for client connections. The - default value is 21050. - :type port: any - :param authentication_type: Required. The authentication type to use. Possible values include: - "Anonymous", "SASLUsername", "UsernameAndPassword". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.ImpalaAuthenticationType - :param username: The user name used to access the Impala server. The default value is anonymous - when using SASLUsername. - :type username: any - :param password: The password corresponding to the user name when using UsernameAndPassword. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The - default value is false. - :type enable_ssl: any - :param trusted_cert_path: The full path of the .pem file containing trusted CA certificates for - verifying the server when connecting over SSL. This property can only be set when using SSL on - self-hosted IR. The default value is the cacerts.pem file installed with the IR. - :type trusted_cert_path: any - :param use_system_trust_store: Specifies whether to use a CA certificate from the system trust - store or from a specified PEM file. The default value is false. - :type use_system_trust_store: any - :param allow_host_name_cn_mismatch: Specifies whether to require a CA-issued SSL certificate - name to match the host name of the server when connecting over SSL. The default value is false. - :type allow_host_name_cn_mismatch: any - :param allow_self_signed_server_cert: Specifies whether to allow self-signed certificates from - the server. The default value is false. - :type allow_self_signed_server_cert: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, - 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, - 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, - 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, - 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ImpalaLinkedService, self).__init__(**kwargs) - self.type = 'Impala' # type: str - self.host = kwargs['host'] - self.port = kwargs.get('port', None) - self.authentication_type = kwargs['authentication_type'] - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.enable_ssl = kwargs.get('enable_ssl', None) - self.trusted_cert_path = kwargs.get('trusted_cert_path', None) - self.use_system_trust_store = kwargs.get('use_system_trust_store', None) - self.allow_host_name_cn_mismatch = kwargs.get('allow_host_name_cn_mismatch', None) - self.allow_self_signed_server_cert = kwargs.get('allow_self_signed_server_cert', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class ImpalaObjectDataset(Dataset): - """Impala server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The table name of the Impala. Type: string (or Expression with resultType - string). - :type table: any - :param schema_type_properties_schema: The schema name of the Impala. Type: string (or - Expression with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ImpalaObjectDataset, self).__init__(**kwargs) - self.type = 'ImpalaObject' # type: str - self.table_name = kwargs.get('table_name', None) - self.table = kwargs.get('table', None) - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - - -class ImpalaSource(TabularSource): - """A copy activity Impala server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ImpalaSource, self).__init__(**kwargs) - self.type = 'ImpalaSource' # type: str - self.query = kwargs.get('query', None) - - -class InformixLinkedService(LinkedService): - """Informix linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The non-access credential portion of the connection string - as well as an optional encrypted credential. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param authentication_type: Type of authentication used to connect to the Informix as ODBC data - store. Possible values are: Anonymous and Basic. Type: string (or Expression with resultType - string). - :type authentication_type: any - :param credential: The access credential portion of the connection string specified in - driver-specific property-value format. - :type credential: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param user_name: User name for Basic authentication. Type: string (or Expression with - resultType string). - :type user_name: any - :param password: Password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, - 'credential': {'key': 'typeProperties.credential', 'type': 'SecretBase'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(InformixLinkedService, self).__init__(**kwargs) - self.type = 'Informix' # type: str - self.connection_string = kwargs['connection_string'] - self.authentication_type = kwargs.get('authentication_type', None) - self.credential = kwargs.get('credential', None) - self.user_name = kwargs.get('user_name', None) - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class InformixSink(CopySink): - """A copy activity Informix sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param pre_copy_script: A query to execute before starting the copy. Type: string (or - Expression with resultType string). - :type pre_copy_script: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(InformixSink, self).__init__(**kwargs) - self.type = 'InformixSink' # type: str - self.pre_copy_script = kwargs.get('pre_copy_script', None) - - -class InformixSource(TabularSource): - """A copy activity source for Informix. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(InformixSource, self).__init__(**kwargs) - self.type = 'InformixSource' # type: str - self.query = kwargs.get('query', None) - - -class InformixTableDataset(Dataset): - """The Informix table dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The Informix table name. Type: string (or Expression with resultType - string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(InformixTableDataset, self).__init__(**kwargs) - self.type = 'InformixTable' # type: str - self.table_name = kwargs.get('table_name', None) - - -class IntegrationRuntime(msrest.serialization.Model): - """Azure Synapse nested object which serves as a compute resource for activities. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: ManagedIntegrationRuntime, SelfHostedIntegrationRuntime. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of integration runtime.Constant filled by server. Possible values - include: "Managed", "SelfHosted". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeType - :param description: Integration runtime description. - :type description: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - } - - _subtype_map = { - 'type': {'Managed': 'ManagedIntegrationRuntime', 'SelfHosted': 'SelfHostedIntegrationRuntime'} - } - - def __init__( - self, - **kwargs - ): - super(IntegrationRuntime, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = 'IntegrationRuntime' # type: str - self.description = kwargs.get('description', None) - - -class IntegrationRuntimeComputeProperties(msrest.serialization.Model): - """The compute resource properties for managed integration runtime. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param location: The location for managed integration runtime. The supported regions could be - found on - https://docs.microsoft.com/en-us/azure/data-factory/data-factory-data-movement-activities. - :type location: str - :param node_size: The node size requirement to managed integration runtime. - :type node_size: str - :param number_of_nodes: The required number of nodes for managed integration runtime. - :type number_of_nodes: int - :param max_parallel_executions_per_node: Maximum parallel executions count per node for managed - integration runtime. - :type max_parallel_executions_per_node: int - :param data_flow_properties: Data flow properties for managed integration runtime. - :type data_flow_properties: - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeDataFlowProperties - :param v_net_properties: VNet properties for managed integration runtime. - :type v_net_properties: - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeVNetProperties - """ - - _validation = { - 'number_of_nodes': {'minimum': 1}, - 'max_parallel_executions_per_node': {'minimum': 1}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'location': {'key': 'location', 'type': 'str'}, - 'node_size': {'key': 'nodeSize', 'type': 'str'}, - 'number_of_nodes': {'key': 'numberOfNodes', 'type': 'int'}, - 'max_parallel_executions_per_node': {'key': 'maxParallelExecutionsPerNode', 'type': 'int'}, - 'data_flow_properties': {'key': 'dataFlowProperties', 'type': 'IntegrationRuntimeDataFlowProperties'}, - 'v_net_properties': {'key': 'vNetProperties', 'type': 'IntegrationRuntimeVNetProperties'}, - } - - def __init__( - self, - **kwargs - ): - super(IntegrationRuntimeComputeProperties, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.location = kwargs.get('location', None) - self.node_size = kwargs.get('node_size', None) - self.number_of_nodes = kwargs.get('number_of_nodes', None) - self.max_parallel_executions_per_node = kwargs.get('max_parallel_executions_per_node', None) - self.data_flow_properties = kwargs.get('data_flow_properties', None) - self.v_net_properties = kwargs.get('v_net_properties', None) - - -class IntegrationRuntimeCustomSetupScriptProperties(msrest.serialization.Model): - """Custom setup script properties for a managed dedicated integration runtime. - - :param blob_container_uri: The URI of the Azure blob container that contains the custom setup - script. - :type blob_container_uri: str - :param sas_token: The SAS token of the Azure blob container. - :type sas_token: ~azure.synapse.artifacts.v2020_12_01.models.SecureString - """ - - _attribute_map = { - 'blob_container_uri': {'key': 'blobContainerUri', 'type': 'str'}, - 'sas_token': {'key': 'sasToken', 'type': 'SecureString'}, - } - - def __init__( - self, - **kwargs - ): - super(IntegrationRuntimeCustomSetupScriptProperties, self).__init__(**kwargs) - self.blob_container_uri = kwargs.get('blob_container_uri', None) - self.sas_token = kwargs.get('sas_token', None) - - -class IntegrationRuntimeDataFlowProperties(msrest.serialization.Model): - """Data flow properties for managed integration runtime. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param compute_type: Compute type of the cluster which will execute data flow job. Possible - values include: "General", "MemoryOptimized", "ComputeOptimized". - :type compute_type: str or ~azure.synapse.artifacts.v2020_12_01.models.DataFlowComputeType - :param core_count: Core count of the cluster which will execute data flow job. Supported values - are: 8, 16, 32, 48, 80, 144 and 272. - :type core_count: int - :param time_to_live: Time to live (in minutes) setting of the cluster which will execute data - flow job. - :type time_to_live: int - """ - - _validation = { - 'time_to_live': {'minimum': 0}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'compute_type': {'key': 'computeType', 'type': 'str'}, - 'core_count': {'key': 'coreCount', 'type': 'int'}, - 'time_to_live': {'key': 'timeToLive', 'type': 'int'}, - } - - def __init__( - self, - **kwargs - ): - super(IntegrationRuntimeDataFlowProperties, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.compute_type = kwargs.get('compute_type', None) - self.core_count = kwargs.get('core_count', None) - self.time_to_live = kwargs.get('time_to_live', None) - - -class IntegrationRuntimeDataProxyProperties(msrest.serialization.Model): - """Data proxy properties for a managed dedicated integration runtime. - - :param connect_via: The self-hosted integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.EntityReference - :param staging_linked_service: The staging linked service reference. - :type staging_linked_service: ~azure.synapse.artifacts.v2020_12_01.models.EntityReference - :param path: The path to contain the staged data in the Blob storage. - :type path: str - """ - - _attribute_map = { - 'connect_via': {'key': 'connectVia', 'type': 'EntityReference'}, - 'staging_linked_service': {'key': 'stagingLinkedService', 'type': 'EntityReference'}, - 'path': {'key': 'path', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(IntegrationRuntimeDataProxyProperties, self).__init__(**kwargs) - self.connect_via = kwargs.get('connect_via', None) - self.staging_linked_service = kwargs.get('staging_linked_service', None) - self.path = kwargs.get('path', None) - - -class IntegrationRuntimeListResponse(msrest.serialization.Model): - """A list of integration runtime resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of integration runtimes. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[IntegrationRuntimeResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(IntegrationRuntimeListResponse, self).__init__(**kwargs) - self.value = kwargs['value'] - self.next_link = kwargs.get('next_link', None) - - -class IntegrationRuntimeReference(msrest.serialization.Model): - """Integration runtime reference type. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Type of integration runtime. Possible values include: - "IntegrationRuntimeReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReferenceType - :param reference_name: Required. Reference integration runtime name. - :type reference_name: str - :param parameters: Arguments for integration runtime. - :type parameters: dict[str, any] - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{object}'}, - } - - def __init__( - self, - **kwargs - ): - super(IntegrationRuntimeReference, self).__init__(**kwargs) - self.type = kwargs['type'] - self.reference_name = kwargs['reference_name'] - self.parameters = kwargs.get('parameters', None) - - -class IntegrationRuntimeResource(SubResource): - """Integration runtime resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Integration runtime properties. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntime - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'IntegrationRuntime'}, - } - - def __init__( - self, - **kwargs - ): - super(IntegrationRuntimeResource, self).__init__(**kwargs) - self.properties = kwargs['properties'] - - -class IntegrationRuntimeSsisCatalogInfo(msrest.serialization.Model): - """Catalog information for managed dedicated integration runtime. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param catalog_server_endpoint: The catalog database server URL. - :type catalog_server_endpoint: str - :param catalog_admin_user_name: The administrator user name of catalog database. - :type catalog_admin_user_name: str - :param catalog_admin_password: The password of the administrator user account of the catalog - database. - :type catalog_admin_password: ~azure.synapse.artifacts.v2020_12_01.models.SecureString - :param catalog_pricing_tier: The pricing tier for the catalog database. The valid values could - be found in https://azure.microsoft.com/en-us/pricing/details/sql-database/. Possible values - include: "Basic", "Standard", "Premium", "PremiumRS". - :type catalog_pricing_tier: str or - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeSsisCatalogPricingTier - """ - - _validation = { - 'catalog_admin_user_name': {'max_length': 128, 'min_length': 1}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'catalog_server_endpoint': {'key': 'catalogServerEndpoint', 'type': 'str'}, - 'catalog_admin_user_name': {'key': 'catalogAdminUserName', 'type': 'str'}, - 'catalog_admin_password': {'key': 'catalogAdminPassword', 'type': 'SecureString'}, - 'catalog_pricing_tier': {'key': 'catalogPricingTier', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(IntegrationRuntimeSsisCatalogInfo, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.catalog_server_endpoint = kwargs.get('catalog_server_endpoint', None) - self.catalog_admin_user_name = kwargs.get('catalog_admin_user_name', None) - self.catalog_admin_password = kwargs.get('catalog_admin_password', None) - self.catalog_pricing_tier = kwargs.get('catalog_pricing_tier', None) - - -class IntegrationRuntimeSsisProperties(msrest.serialization.Model): - """SSIS properties for managed integration runtime. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param catalog_info: Catalog information for managed dedicated integration runtime. - :type catalog_info: - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeSsisCatalogInfo - :param license_type: License type for bringing your own license scenario. Possible values - include: "BasePrice", "LicenseIncluded". - :type license_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeLicenseType - :param custom_setup_script_properties: Custom setup script properties for a managed dedicated - integration runtime. - :type custom_setup_script_properties: - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeCustomSetupScriptProperties - :param data_proxy_properties: Data proxy properties for a managed dedicated integration - runtime. - :type data_proxy_properties: - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeDataProxyProperties - :param edition: The edition for the SSIS Integration Runtime. Possible values include: - "Standard", "Enterprise". - :type edition: str or ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeEdition - :param express_custom_setup_properties: Custom setup without script properties for a SSIS - integration runtime. - :type express_custom_setup_properties: - list[~azure.synapse.artifacts.v2020_12_01.models.CustomSetupBase] - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'catalog_info': {'key': 'catalogInfo', 'type': 'IntegrationRuntimeSsisCatalogInfo'}, - 'license_type': {'key': 'licenseType', 'type': 'str'}, - 'custom_setup_script_properties': {'key': 'customSetupScriptProperties', 'type': 'IntegrationRuntimeCustomSetupScriptProperties'}, - 'data_proxy_properties': {'key': 'dataProxyProperties', 'type': 'IntegrationRuntimeDataProxyProperties'}, - 'edition': {'key': 'edition', 'type': 'str'}, - 'express_custom_setup_properties': {'key': 'expressCustomSetupProperties', 'type': '[CustomSetupBase]'}, - } - - def __init__( - self, - **kwargs - ): - super(IntegrationRuntimeSsisProperties, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.catalog_info = kwargs.get('catalog_info', None) - self.license_type = kwargs.get('license_type', None) - self.custom_setup_script_properties = kwargs.get('custom_setup_script_properties', None) - self.data_proxy_properties = kwargs.get('data_proxy_properties', None) - self.edition = kwargs.get('edition', None) - self.express_custom_setup_properties = kwargs.get('express_custom_setup_properties', None) - - -class IntegrationRuntimeVNetProperties(msrest.serialization.Model): - """VNet properties for managed integration runtime. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param v_net_id: The ID of the VNet that this integration runtime will join. - :type v_net_id: str - :param subnet: The name of the subnet this integration runtime will join. - :type subnet: str - :param public_i_ps: Resource IDs of the public IP addresses that this integration runtime will - use. - :type public_i_ps: list[str] - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'v_net_id': {'key': 'vNetId', 'type': 'str'}, - 'subnet': {'key': 'subnet', 'type': 'str'}, - 'public_i_ps': {'key': 'publicIPs', 'type': '[str]'}, - } - - def __init__( - self, - **kwargs - ): - super(IntegrationRuntimeVNetProperties, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.v_net_id = kwargs.get('v_net_id', None) - self.subnet = kwargs.get('subnet', None) - self.public_i_ps = kwargs.get('public_i_ps', None) - - -class JiraLinkedService(LinkedService): - """Jira Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The IP address or host name of the Jira service. (e.g. - jira.example.com). - :type host: any - :param port: The TCP port that the Jira server uses to listen for client connections. The - default value is 443 if connecting through HTTPS, or 8080 if connecting through HTTP. - :type port: any - :param username: Required. The user name that you use to access Jira Service. - :type username: any - :param password: The password corresponding to the user name that you provided in the username - field. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'username': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(JiraLinkedService, self).__init__(**kwargs) - self.type = 'Jira' # type: str - self.host = kwargs['host'] - self.port = kwargs.get('port', None) - self.username = kwargs['username'] - self.password = kwargs.get('password', None) - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.use_host_verification = kwargs.get('use_host_verification', None) - self.use_peer_verification = kwargs.get('use_peer_verification', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class JiraObjectDataset(Dataset): - """Jira Service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(JiraObjectDataset, self).__init__(**kwargs) - self.type = 'JiraObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class JiraSource(TabularSource): - """A copy activity Jira Service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(JiraSource, self).__init__(**kwargs) - self.type = 'JiraSource' # type: str - self.query = kwargs.get('query', None) - - -class JsonDataset(Dataset): - """Json dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param location: The location of the json data storage. - :type location: ~azure.synapse.artifacts.v2020_12_01.models.DatasetLocation - :param encoding_name: The code page name of the preferred encoding. If not specified, the - default value is UTF-8, unless BOM denotes another Unicode encoding. Refer to the name column - of the table in the following link to set supported values: - https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string (or Expression with - resultType string). - :type encoding_name: any - :param compression: The data compression method used for the json dataset. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'location': {'key': 'typeProperties.location', 'type': 'DatasetLocation'}, - 'encoding_name': {'key': 'typeProperties.encodingName', 'type': 'object'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - } - - def __init__( - self, - **kwargs - ): - super(JsonDataset, self).__init__(**kwargs) - self.type = 'Json' # type: str - self.location = kwargs.get('location', None) - self.encoding_name = kwargs.get('encoding_name', None) - self.compression = kwargs.get('compression', None) - - -class JsonFormat(DatasetStorageFormat): - """The data stored in JSON format. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage format.Constant filled by server. - :type type: str - :param serializer: Serializer. Type: string (or Expression with resultType string). - :type serializer: any - :param deserializer: Deserializer. Type: string (or Expression with resultType string). - :type deserializer: any - :param file_pattern: File pattern of JSON. To be more specific, the way of separating a - collection of JSON objects. The default value is 'setOfObjects'. It is case-sensitive. Possible - values include: "setOfObjects", "arrayOfObjects". - :type file_pattern: str or ~azure.synapse.artifacts.v2020_12_01.models.JsonFormatFilePattern - :param nesting_separator: The character used to separate nesting levels. Default value is '.' - (dot). Type: string (or Expression with resultType string). - :type nesting_separator: any - :param encoding_name: The code page name of the preferred encoding. If not provided, the - default value is 'utf-8', unless the byte order mark (BOM) denotes another Unicode encoding. - The full list of supported values can be found in the 'Name' column of the table of encodings - in the following reference: https://go.microsoft.com/fwlink/?linkid=861078. Type: string (or - Expression with resultType string). - :type encoding_name: any - :param json_node_reference: The JSONPath of the JSON array element to be flattened. Example: - "$.ArrayPath". Type: string (or Expression with resultType string). - :type json_node_reference: any - :param json_path_definition: The JSONPath definition for each column mapping with a customized - column name to extract data from JSON file. For fields under root object, start with "$"; for - fields inside the array chosen by jsonNodeReference property, start from the array element. - Example: {"Column1": "$.Column1Path", "Column2": "Column2PathInArray"}. Type: object (or - Expression with resultType object). - :type json_path_definition: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'serializer': {'key': 'serializer', 'type': 'object'}, - 'deserializer': {'key': 'deserializer', 'type': 'object'}, - 'file_pattern': {'key': 'filePattern', 'type': 'str'}, - 'nesting_separator': {'key': 'nestingSeparator', 'type': 'object'}, - 'encoding_name': {'key': 'encodingName', 'type': 'object'}, - 'json_node_reference': {'key': 'jsonNodeReference', 'type': 'object'}, - 'json_path_definition': {'key': 'jsonPathDefinition', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(JsonFormat, self).__init__(**kwargs) - self.type = 'JsonFormat' # type: str - self.file_pattern = kwargs.get('file_pattern', None) - self.nesting_separator = kwargs.get('nesting_separator', None) - self.encoding_name = kwargs.get('encoding_name', None) - self.json_node_reference = kwargs.get('json_node_reference', None) - self.json_path_definition = kwargs.get('json_path_definition', None) - - -class JsonReadSettings(FormatReadSettings): - """Json read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param compression_properties: Compression settings. - :type compression_properties: - ~azure.synapse.artifacts.v2020_12_01.models.CompressionReadSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'compression_properties': {'key': 'compressionProperties', 'type': 'CompressionReadSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(JsonReadSettings, self).__init__(**kwargs) - self.type = 'JsonReadSettings' # type: str - self.compression_properties = kwargs.get('compression_properties', None) - - -class JsonSink(CopySink): - """A copy activity Json sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Json store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreWriteSettings - :param format_settings: Json format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.JsonWriteSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreWriteSettings'}, - 'format_settings': {'key': 'formatSettings', 'type': 'JsonWriteSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(JsonSink, self).__init__(**kwargs) - self.type = 'JsonSink' # type: str - self.store_settings = kwargs.get('store_settings', None) - self.format_settings = kwargs.get('format_settings', None) - - -class JsonSource(CopySource): - """A copy activity Json source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Json store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - :param format_settings: Json format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.JsonReadSettings - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'format_settings': {'key': 'formatSettings', 'type': 'JsonReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(JsonSource, self).__init__(**kwargs) - self.type = 'JsonSource' # type: str - self.store_settings = kwargs.get('store_settings', None) - self.format_settings = kwargs.get('format_settings', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class JsonWriteSettings(FormatWriteSettings): - """Json write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param file_pattern: File pattern of JSON. This setting controls the way a collection of JSON - objects will be treated. The default value is 'setOfObjects'. It is case-sensitive. Possible - values include: "setOfObjects", "arrayOfObjects". - :type file_pattern: str or ~azure.synapse.artifacts.v2020_12_01.models.JsonWriteFilePattern - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'file_pattern': {'key': 'filePattern', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(JsonWriteSettings, self).__init__(**kwargs) - self.type = 'JsonWriteSettings' # type: str - self.file_pattern = kwargs.get('file_pattern', None) - - -class LibraryInfo(msrest.serialization.Model): - """Library/package information of a Big Data pool powered by Apache Spark. - - Variables are only populated by the server, and will be ignored when sending a request. - - :param name: Name of the library. - :type name: str - :param path: Storage blob path of library. - :type path: str - :param container_name: Storage blob container name. - :type container_name: str - :ivar uploaded_timestamp: The last update time of the library. - :vartype uploaded_timestamp: ~datetime.datetime - :param type: Type of the library. - :type type: str - :ivar provisioning_status: Provisioning status of the library/package. - :vartype provisioning_status: str - :ivar creator_id: Creator Id of the library/package. - :vartype creator_id: str - """ - - _validation = { - 'uploaded_timestamp': {'readonly': True}, - 'provisioning_status': {'readonly': True}, - 'creator_id': {'readonly': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'path': {'key': 'path', 'type': 'str'}, - 'container_name': {'key': 'containerName', 'type': 'str'}, - 'uploaded_timestamp': {'key': 'uploadedTimestamp', 'type': 'iso-8601'}, - 'type': {'key': 'type', 'type': 'str'}, - 'provisioning_status': {'key': 'provisioningStatus', 'type': 'str'}, - 'creator_id': {'key': 'creatorId', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(LibraryInfo, self).__init__(**kwargs) - self.name = kwargs.get('name', None) - self.path = kwargs.get('path', None) - self.container_name = kwargs.get('container_name', None) - self.uploaded_timestamp = None - self.type = kwargs.get('type', None) - self.provisioning_status = None - self.creator_id = None - - -class LibraryListResponse(msrest.serialization.Model): - """A list of Library resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of Library. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.LibraryResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[LibraryResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(LibraryListResponse, self).__init__(**kwargs) - self.value = kwargs['value'] - self.next_link = kwargs.get('next_link', None) - - -class LibraryRequirements(msrest.serialization.Model): - """Library requirements for a Big Data pool powered by Apache Spark. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar time: The last update time of the library requirements file. - :vartype time: ~datetime.datetime - :param content: The library requirements. - :type content: str - :param filename: The filename of the library requirements file. - :type filename: str - """ - - _validation = { - 'time': {'readonly': True}, - } - - _attribute_map = { - 'time': {'key': 'time', 'type': 'iso-8601'}, - 'content': {'key': 'content', 'type': 'str'}, - 'filename': {'key': 'filename', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(LibraryRequirements, self).__init__(**kwargs) - self.time = None - self.content = kwargs.get('content', None) - self.filename = kwargs.get('filename', None) - - -class LibraryResource(SubResource): - """Library response details. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Library/package properties. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.LibraryResourceProperties - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'LibraryResourceProperties'}, - } - - def __init__( - self, - **kwargs - ): - super(LibraryResource, self).__init__(**kwargs) - self.properties = kwargs['properties'] - - -class LibraryResourceInfo(msrest.serialization.Model): - """Library resource info. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar record_id: record Id of the library/package. - :vartype record_id: int - :ivar state: Provisioning status of the library/package. - :vartype state: str - :ivar created: The creation time of the library/package. - :vartype created: str - :ivar changed: The last updated time of the library/package. - :vartype changed: str - :ivar type: The type of the resource. E.g. LibraryArtifact. - :vartype type: str - :ivar name: Name of the library/package. - :vartype name: str - :ivar operation_id: Operation Id of the operation performed on library/package. - :vartype operation_id: str - :ivar artifact_id: artifact Id of the library/package. - :vartype artifact_id: str - """ - - _validation = { - 'id': {'readonly': True}, - 'record_id': {'readonly': True}, - 'state': {'readonly': True}, - 'created': {'readonly': True}, - 'changed': {'readonly': True}, - 'type': {'readonly': True}, - 'name': {'readonly': True}, - 'operation_id': {'readonly': True}, - 'artifact_id': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'record_id': {'key': 'recordId', 'type': 'int'}, - 'state': {'key': 'state', 'type': 'str'}, - 'created': {'key': 'created', 'type': 'str'}, - 'changed': {'key': 'changed', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'operation_id': {'key': 'operationId', 'type': 'str'}, - 'artifact_id': {'key': 'artifactId', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(LibraryResourceInfo, self).__init__(**kwargs) - self.id = None - self.record_id = None - self.state = None - self.created = None - self.changed = None - self.type = None - self.name = None - self.operation_id = None - self.artifact_id = None - - -class LibraryResourceProperties(msrest.serialization.Model): - """Library/package properties. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar name: Name of the library/package. - :vartype name: str - :ivar path: Location of library/package in storage account. - :vartype path: str - :ivar container_name: Container name of the library/package. - :vartype container_name: str - :ivar uploaded_timestamp: The last update time of the library/package. - :vartype uploaded_timestamp: str - :ivar type: Type of the library/package. - :vartype type: str - :ivar provisioning_status: Provisioning status of the library/package. - :vartype provisioning_status: str - :ivar creator_id: Creator Id of the library/package. - :vartype creator_id: str - """ - - _validation = { - 'name': {'readonly': True}, - 'path': {'readonly': True}, - 'container_name': {'readonly': True}, - 'uploaded_timestamp': {'readonly': True}, - 'type': {'readonly': True}, - 'provisioning_status': {'readonly': True}, - 'creator_id': {'readonly': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'path': {'key': 'path', 'type': 'str'}, - 'container_name': {'key': 'containerName', 'type': 'str'}, - 'uploaded_timestamp': {'key': 'uploadedTimestamp', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'provisioning_status': {'key': 'provisioningStatus', 'type': 'str'}, - 'creator_id': {'key': 'creatorId', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(LibraryResourceProperties, self).__init__(**kwargs) - self.name = None - self.path = None - self.container_name = None - self.uploaded_timestamp = None - self.type = None - self.provisioning_status = None - self.creator_id = None - - -class LinkedIntegrationRuntimeType(msrest.serialization.Model): - """The base definition of a linked integration runtime. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: LinkedIntegrationRuntimeKeyAuthorization, LinkedIntegrationRuntimeRbacAuthorization. - - All required parameters must be populated in order to send to Azure. - - :param authorization_type: Required. The authorization type for integration runtime - sharing.Constant filled by server. - :type authorization_type: str - """ - - _validation = { - 'authorization_type': {'required': True}, - } - - _attribute_map = { - 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, - } - - _subtype_map = { - 'authorization_type': {'Key': 'LinkedIntegrationRuntimeKeyAuthorization', 'RBAC': 'LinkedIntegrationRuntimeRbacAuthorization'} - } - - def __init__( - self, - **kwargs - ): - super(LinkedIntegrationRuntimeType, self).__init__(**kwargs) - self.authorization_type = None # type: Optional[str] - - -class LinkedIntegrationRuntimeKeyAuthorization(LinkedIntegrationRuntimeType): - """The key authorization type integration runtime. - - All required parameters must be populated in order to send to Azure. - - :param authorization_type: Required. The authorization type for integration runtime - sharing.Constant filled by server. - :type authorization_type: str - :param key: Required. The key used for authorization. - :type key: ~azure.synapse.artifacts.v2020_12_01.models.SecureString - """ - - _validation = { - 'authorization_type': {'required': True}, - 'key': {'required': True}, - } - - _attribute_map = { - 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, - 'key': {'key': 'key', 'type': 'SecureString'}, - } - - def __init__( - self, - **kwargs - ): - super(LinkedIntegrationRuntimeKeyAuthorization, self).__init__(**kwargs) - self.authorization_type = 'Key' # type: str - self.key = kwargs['key'] - - -class LinkedIntegrationRuntimeRbacAuthorization(LinkedIntegrationRuntimeType): - """The role based access control (RBAC) authorization type integration runtime. - - All required parameters must be populated in order to send to Azure. - - :param authorization_type: Required. The authorization type for integration runtime - sharing.Constant filled by server. - :type authorization_type: str - :param resource_id: Required. The resource identifier of the integration runtime to be shared. - :type resource_id: str - """ - - _validation = { - 'authorization_type': {'required': True}, - 'resource_id': {'required': True}, - } - - _attribute_map = { - 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, - 'resource_id': {'key': 'resourceId', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(LinkedIntegrationRuntimeRbacAuthorization, self).__init__(**kwargs) - self.authorization_type = 'RBAC' # type: str - self.resource_id = kwargs['resource_id'] - - -class LinkedServiceDebugResource(SubResourceDebugResource): - """Linked service debug resource. - - All required parameters must be populated in order to send to Azure. - - :param name: The resource name. - :type name: str - :param properties: Required. Properties of linked service. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.LinkedService - """ - - _validation = { - 'properties': {'required': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'LinkedService'}, - } - - def __init__( - self, - **kwargs - ): - super(LinkedServiceDebugResource, self).__init__(**kwargs) - self.properties = kwargs['properties'] - - -class LinkedServiceListResponse(msrest.serialization.Model): - """A list of linked service resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of linked services. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[LinkedServiceResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(LinkedServiceListResponse, self).__init__(**kwargs) - self.value = kwargs['value'] - self.next_link = kwargs.get('next_link', None) - - -class LinkedServiceReference(msrest.serialization.Model): - """Linked service reference type. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Linked service reference type. Possible values include: - "LinkedServiceReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.Type - :param reference_name: Required. Reference LinkedService name. - :type reference_name: str - :param parameters: Arguments for LinkedService. - :type parameters: dict[str, any] - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{object}'}, - } - - def __init__( - self, - **kwargs - ): - super(LinkedServiceReference, self).__init__(**kwargs) - self.type = kwargs['type'] - self.reference_name = kwargs['reference_name'] - self.parameters = kwargs.get('parameters', None) - - -class LinkedServiceResource(SubResource): - """Linked service resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Properties of linked service. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.LinkedService - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'LinkedService'}, - } - - def __init__( - self, - **kwargs - ): - super(LinkedServiceResource, self).__init__(**kwargs) - self.properties = kwargs['properties'] - - -class LogLocationSettings(msrest.serialization.Model): - """Log location settings. - - All required parameters must be populated in order to send to Azure. - - :param linked_service_name: Required. Log storage linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param path: The path to storage for storing detailed logs of activity execution. Type: string - (or Expression with resultType string). - :type path: any - """ - - _validation = { - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'path': {'key': 'path', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(LogLocationSettings, self).__init__(**kwargs) - self.linked_service_name = kwargs['linked_service_name'] - self.path = kwargs.get('path', None) - - -class LogSettings(msrest.serialization.Model): - """Log settings. - - All required parameters must be populated in order to send to Azure. - - :param enable_copy_activity_log: Specifies whether to enable copy activity log. Type: boolean - (or Expression with resultType boolean). - :type enable_copy_activity_log: any - :param copy_activity_log_settings: Specifies settings for copy activity log. - :type copy_activity_log_settings: - ~azure.synapse.artifacts.v2020_12_01.models.CopyActivityLogSettings - :param log_location_settings: Required. Log location settings customer needs to provide when - enabling log. - :type log_location_settings: ~azure.synapse.artifacts.v2020_12_01.models.LogLocationSettings - """ - - _validation = { - 'log_location_settings': {'required': True}, - } - - _attribute_map = { - 'enable_copy_activity_log': {'key': 'enableCopyActivityLog', 'type': 'object'}, - 'copy_activity_log_settings': {'key': 'copyActivityLogSettings', 'type': 'CopyActivityLogSettings'}, - 'log_location_settings': {'key': 'logLocationSettings', 'type': 'LogLocationSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(LogSettings, self).__init__(**kwargs) - self.enable_copy_activity_log = kwargs.get('enable_copy_activity_log', None) - self.copy_activity_log_settings = kwargs.get('copy_activity_log_settings', None) - self.log_location_settings = kwargs['log_location_settings'] - - -class LogStorageSettings(msrest.serialization.Model): - """(Deprecated. Please use LogSettings) Log storage settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param linked_service_name: Required. Log storage linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param path: The path to storage for storing detailed logs of activity execution. Type: string - (or Expression with resultType string). - :type path: any - :param log_level: Gets or sets the log level, support: Info, Warning. Type: string (or - Expression with resultType string). - :type log_level: any - :param enable_reliable_logging: Specifies whether to enable reliable logging. Type: boolean (or - Expression with resultType boolean). - :type enable_reliable_logging: any - """ - - _validation = { - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'path': {'key': 'path', 'type': 'object'}, - 'log_level': {'key': 'logLevel', 'type': 'object'}, - 'enable_reliable_logging': {'key': 'enableReliableLogging', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(LogStorageSettings, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.linked_service_name = kwargs['linked_service_name'] - self.path = kwargs.get('path', None) - self.log_level = kwargs.get('log_level', None) - self.enable_reliable_logging = kwargs.get('enable_reliable_logging', None) - - -class LookupActivity(ExecutionActivity): - """Lookup activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param source: Required. Dataset-specific source properties, same as copy activity source. - :type source: ~azure.synapse.artifacts.v2020_12_01.models.CopySource - :param dataset: Required. Lookup activity dataset reference. - :type dataset: ~azure.synapse.artifacts.v2020_12_01.models.DatasetReference - :param first_row_only: Whether to return first row or all rows. Default value is true. Type: - boolean (or Expression with resultType boolean). - :type first_row_only: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'source': {'required': True}, - 'dataset': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'source': {'key': 'typeProperties.source', 'type': 'CopySource'}, - 'dataset': {'key': 'typeProperties.dataset', 'type': 'DatasetReference'}, - 'first_row_only': {'key': 'typeProperties.firstRowOnly', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(LookupActivity, self).__init__(**kwargs) - self.type = 'Lookup' # type: str - self.source = kwargs['source'] - self.dataset = kwargs['dataset'] - self.first_row_only = kwargs.get('first_row_only', None) - - -class MagentoLinkedService(LinkedService): - """Magento server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The URL of the Magento instance. (i.e. 192.168.222.110/magento3). - :type host: any - :param access_token: The access token from Magento. - :type access_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MagentoLinkedService, self).__init__(**kwargs) - self.type = 'Magento' # type: str - self.host = kwargs['host'] - self.access_token = kwargs.get('access_token', None) - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.use_host_verification = kwargs.get('use_host_verification', None) - self.use_peer_verification = kwargs.get('use_peer_verification', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class MagentoObjectDataset(Dataset): - """Magento server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MagentoObjectDataset, self).__init__(**kwargs) - self.type = 'MagentoObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class MagentoSource(TabularSource): - """A copy activity Magento server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MagentoSource, self).__init__(**kwargs) - self.type = 'MagentoSource' # type: str - self.query = kwargs.get('query', None) - - -class ManagedIdentity(msrest.serialization.Model): - """The workspace managed identity. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar principal_id: The principal ID of the workspace managed identity. - :vartype principal_id: str - :ivar tenant_id: The tenant ID of the workspace managed identity. - :vartype tenant_id: str - :param type: The type of managed identity for the workspace. Possible values include: "None", - "SystemAssigned". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.ResourceIdentityType - """ - - _validation = { - 'principal_id': {'readonly': True}, - 'tenant_id': {'readonly': True}, - } - - _attribute_map = { - 'principal_id': {'key': 'principalId', 'type': 'str'}, - 'tenant_id': {'key': 'tenantId', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(ManagedIdentity, self).__init__(**kwargs) - self.principal_id = None - self.tenant_id = None - self.type = kwargs.get('type', None) - - -class ManagedIntegrationRuntime(IntegrationRuntime): - """Managed integration runtime, including managed elastic and managed dedicated integration runtimes. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of integration runtime.Constant filled by server. Possible values - include: "Managed", "SelfHosted". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeType - :param description: Integration runtime description. - :type description: str - :ivar state: Integration runtime state, only valid for managed dedicated integration runtime. - Possible values include: "Initial", "Stopped", "Started", "Starting", "Stopping", - "NeedRegistration", "Online", "Limited", "Offline", "AccessDenied". - :vartype state: str or ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeState - :param managed_virtual_network: Managed Virtual Network reference. - :type managed_virtual_network: - ~azure.synapse.artifacts.v2020_12_01.models.ManagedVirtualNetworkReference - :param compute_properties: The compute resource for managed integration runtime. - :type compute_properties: - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeComputeProperties - :param ssis_properties: SSIS properties for managed integration runtime. - :type ssis_properties: - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeSsisProperties - """ - - _validation = { - 'type': {'required': True}, - 'state': {'readonly': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'state': {'key': 'state', 'type': 'str'}, - 'managed_virtual_network': {'key': 'managedVirtualNetwork', 'type': 'ManagedVirtualNetworkReference'}, - 'compute_properties': {'key': 'typeProperties.computeProperties', 'type': 'IntegrationRuntimeComputeProperties'}, - 'ssis_properties': {'key': 'typeProperties.ssisProperties', 'type': 'IntegrationRuntimeSsisProperties'}, - } - - def __init__( - self, - **kwargs - ): - super(ManagedIntegrationRuntime, self).__init__(**kwargs) - self.type = 'Managed' # type: str - self.state = None - self.managed_virtual_network = kwargs.get('managed_virtual_network', None) - self.compute_properties = kwargs.get('compute_properties', None) - self.ssis_properties = kwargs.get('ssis_properties', None) - - -class ManagedVirtualNetworkReference(msrest.serialization.Model): - """Managed Virtual Network reference type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar type: Managed Virtual Network reference type. Has constant value: - "ManagedVirtualNetworkReference". - :vartype type: str - :param reference_name: Required. Reference ManagedVirtualNetwork name. - :type reference_name: str - """ - - _validation = { - 'type': {'required': True, 'constant': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - } - - type = "ManagedVirtualNetworkReference" - - def __init__( - self, - **kwargs - ): - super(ManagedVirtualNetworkReference, self).__init__(**kwargs) - self.reference_name = kwargs['reference_name'] - - -class ManagedVirtualNetworkSettings(msrest.serialization.Model): - """Managed Virtual Network Settings. - - :param prevent_data_exfiltration: Prevent Data Exfiltration. - :type prevent_data_exfiltration: bool - :param linked_access_check_on_target_resource: Linked Access Check On Target Resource. - :type linked_access_check_on_target_resource: bool - :param allowed_aad_tenant_ids_for_linking: Allowed Aad Tenant Ids For Linking. - :type allowed_aad_tenant_ids_for_linking: list[str] - """ - - _attribute_map = { - 'prevent_data_exfiltration': {'key': 'preventDataExfiltration', 'type': 'bool'}, - 'linked_access_check_on_target_resource': {'key': 'linkedAccessCheckOnTargetResource', 'type': 'bool'}, - 'allowed_aad_tenant_ids_for_linking': {'key': 'allowedAadTenantIdsForLinking', 'type': '[str]'}, - } - - def __init__( - self, - **kwargs - ): - super(ManagedVirtualNetworkSettings, self).__init__(**kwargs) - self.prevent_data_exfiltration = kwargs.get('prevent_data_exfiltration', None) - self.linked_access_check_on_target_resource = kwargs.get('linked_access_check_on_target_resource', None) - self.allowed_aad_tenant_ids_for_linking = kwargs.get('allowed_aad_tenant_ids_for_linking', None) - - -class MappingDataFlow(DataFlow): - """Mapping data flow. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Type of data flow.Constant filled by server. - :type type: str - :param description: The description of the data flow. - :type description: str - :param annotations: List of tags that can be used for describing the data flow. - :type annotations: list[any] - :param folder: The folder that this data flow is in. If not specified, Data flow will appear at - the root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowFolder - :param sources: List of sources in data flow. - :type sources: list[~azure.synapse.artifacts.v2020_12_01.models.DataFlowSource] - :param sinks: List of sinks in data flow. - :type sinks: list[~azure.synapse.artifacts.v2020_12_01.models.DataFlowSink] - :param transformations: List of transformations in data flow. - :type transformations: list[~azure.synapse.artifacts.v2020_12_01.models.Transformation] - :param script: DataFlow script. - :type script: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DataFlowFolder'}, - 'sources': {'key': 'typeProperties.sources', 'type': '[DataFlowSource]'}, - 'sinks': {'key': 'typeProperties.sinks', 'type': '[DataFlowSink]'}, - 'transformations': {'key': 'typeProperties.transformations', 'type': '[Transformation]'}, - 'script': {'key': 'typeProperties.script', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(MappingDataFlow, self).__init__(**kwargs) - self.type = 'MappingDataFlow' # type: str - self.sources = kwargs.get('sources', None) - self.sinks = kwargs.get('sinks', None) - self.transformations = kwargs.get('transformations', None) - self.script = kwargs.get('script', None) - - -class MariaDBLinkedService(LinkedService): - """MariaDB server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: An ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MariaDBLinkedService, self).__init__(**kwargs) - self.type = 'MariaDB' # type: str - self.connection_string = kwargs.get('connection_string', None) - self.pwd = kwargs.get('pwd', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class MariaDBSource(TabularSource): - """A copy activity MariaDB server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MariaDBSource, self).__init__(**kwargs) - self.type = 'MariaDBSource' # type: str - self.query = kwargs.get('query', None) - - -class MariaDBTableDataset(Dataset): - """MariaDB server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MariaDBTableDataset, self).__init__(**kwargs) - self.type = 'MariaDBTable' # type: str - self.table_name = kwargs.get('table_name', None) - - -class MarketoLinkedService(LinkedService): - """Marketo server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param endpoint: Required. The endpoint of the Marketo server. (i.e. 123-ABC-321.mktorest.com). - :type endpoint: any - :param client_id: Required. The client Id of your Marketo service. - :type client_id: any - :param client_secret: The client secret of your Marketo service. - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'endpoint': {'required': True}, - 'client_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MarketoLinkedService, self).__init__(**kwargs) - self.type = 'Marketo' # type: str - self.endpoint = kwargs['endpoint'] - self.client_id = kwargs['client_id'] - self.client_secret = kwargs.get('client_secret', None) - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.use_host_verification = kwargs.get('use_host_verification', None) - self.use_peer_verification = kwargs.get('use_peer_verification', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class MarketoObjectDataset(Dataset): - """Marketo server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MarketoObjectDataset, self).__init__(**kwargs) - self.type = 'MarketoObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class MarketoSource(TabularSource): - """A copy activity Marketo server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MarketoSource, self).__init__(**kwargs) - self.type = 'MarketoSource' # type: str - self.query = kwargs.get('query', None) - - -class MicrosoftAccessLinkedService(LinkedService): - """Microsoft Access linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The non-access credential portion of the connection string - as well as an optional encrypted credential. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param authentication_type: Type of authentication used to connect to the Microsoft Access as - ODBC data store. Possible values are: Anonymous and Basic. Type: string (or Expression with - resultType string). - :type authentication_type: any - :param credential: The access credential portion of the connection string specified in - driver-specific property-value format. - :type credential: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param user_name: User name for Basic authentication. Type: string (or Expression with - resultType string). - :type user_name: any - :param password: Password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, - 'credential': {'key': 'typeProperties.credential', 'type': 'SecretBase'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MicrosoftAccessLinkedService, self).__init__(**kwargs) - self.type = 'MicrosoftAccess' # type: str - self.connection_string = kwargs['connection_string'] - self.authentication_type = kwargs.get('authentication_type', None) - self.credential = kwargs.get('credential', None) - self.user_name = kwargs.get('user_name', None) - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class MicrosoftAccessSink(CopySink): - """A copy activity Microsoft Access sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param pre_copy_script: A query to execute before starting the copy. Type: string (or - Expression with resultType string). - :type pre_copy_script: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MicrosoftAccessSink, self).__init__(**kwargs) - self.type = 'MicrosoftAccessSink' # type: str - self.pre_copy_script = kwargs.get('pre_copy_script', None) - - -class MicrosoftAccessSource(CopySource): - """A copy activity source for Microsoft Access. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MicrosoftAccessSource, self).__init__(**kwargs) - self.type = 'MicrosoftAccessSource' # type: str - self.query = kwargs.get('query', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class MicrosoftAccessTableDataset(Dataset): - """The Microsoft Access table dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The Microsoft Access table name. Type: string (or Expression with resultType - string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MicrosoftAccessTableDataset, self).__init__(**kwargs) - self.type = 'MicrosoftAccessTable' # type: str - self.table_name = kwargs.get('table_name', None) - - -class MongoDbAtlasCollectionDataset(Dataset): - """The MongoDB Atlas database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param collection: Required. The collection name of the MongoDB Atlas database. Type: string - (or Expression with resultType string). - :type collection: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'collection': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'collection': {'key': 'typeProperties.collection', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MongoDbAtlasCollectionDataset, self).__init__(**kwargs) - self.type = 'MongoDbAtlasCollection' # type: str - self.collection = kwargs['collection'] - - -class MongoDbAtlasLinkedService(LinkedService): - """Linked service for MongoDB Atlas data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The MongoDB Atlas connection string. Type: string, - SecureString or AzureKeyVaultSecretReference. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param database: Required. The name of the MongoDB Atlas database that you want to access. - Type: string (or Expression with resultType string). - :type database: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - 'database': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MongoDbAtlasLinkedService, self).__init__(**kwargs) - self.type = 'MongoDbAtlas' # type: str - self.connection_string = kwargs['connection_string'] - self.database = kwargs['database'] - - -class MongoDbAtlasSource(CopySource): - """A copy activity source for a MongoDB Atlas database. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param filter: Specifies selection filter using query operators. To return all documents in a - collection, omit this parameter or pass an empty document ({}). Type: string (or Expression - with resultType string). - :type filter: any - :param cursor_methods: Cursor methods for Mongodb query. - :type cursor_methods: - ~azure.synapse.artifacts.v2020_12_01.models.MongoDbCursorMethodsProperties - :param batch_size: Specifies the number of documents to return in each batch of the response - from MongoDB Atlas instance. In most cases, modifying the batch size will not affect the user - or the application. This property's main purpose is to avoid hit the limitation of response - size. Type: integer (or Expression with resultType integer). - :type batch_size: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'filter': {'key': 'filter', 'type': 'object'}, - 'cursor_methods': {'key': 'cursorMethods', 'type': 'MongoDbCursorMethodsProperties'}, - 'batch_size': {'key': 'batchSize', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MongoDbAtlasSource, self).__init__(**kwargs) - self.type = 'MongoDbAtlasSource' # type: str - self.filter = kwargs.get('filter', None) - self.cursor_methods = kwargs.get('cursor_methods', None) - self.batch_size = kwargs.get('batch_size', None) - self.query_timeout = kwargs.get('query_timeout', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class MongoDbCollectionDataset(Dataset): - """The MongoDB database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param collection_name: Required. The table name of the MongoDB database. Type: string (or - Expression with resultType string). - :type collection_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'collection_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'collection_name': {'key': 'typeProperties.collectionName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MongoDbCollectionDataset, self).__init__(**kwargs) - self.type = 'MongoDbCollection' # type: str - self.collection_name = kwargs['collection_name'] - - -class MongoDbCursorMethodsProperties(msrest.serialization.Model): - """Cursor methods for Mongodb query. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param project: Specifies the fields to return in the documents that match the query filter. To - return all fields in the matching documents, omit this parameter. Type: string (or Expression - with resultType string). - :type project: any - :param sort: Specifies the order in which the query returns matching documents. Type: string - (or Expression with resultType string). Type: string (or Expression with resultType string). - :type sort: any - :param skip: Specifies the how many documents skipped and where MongoDB begins returning - results. This approach may be useful in implementing paginated results. Type: integer (or - Expression with resultType integer). - :type skip: any - :param limit: Specifies the maximum number of documents the server returns. limit() is - analogous to the LIMIT statement in a SQL database. Type: integer (or Expression with - resultType integer). - :type limit: any - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'project': {'key': 'project', 'type': 'object'}, - 'sort': {'key': 'sort', 'type': 'object'}, - 'skip': {'key': 'skip', 'type': 'object'}, - 'limit': {'key': 'limit', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MongoDbCursorMethodsProperties, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.project = kwargs.get('project', None) - self.sort = kwargs.get('sort', None) - self.skip = kwargs.get('skip', None) - self.limit = kwargs.get('limit', None) - - -class MongoDbLinkedService(LinkedService): - """Linked service for MongoDb data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param server: Required. The IP address or server name of the MongoDB server. Type: string (or - Expression with resultType string). - :type server: any - :param authentication_type: The authentication type to be used to connect to the MongoDB - database. Possible values include: "Basic", "Anonymous". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.MongoDbAuthenticationType - :param database_name: Required. The name of the MongoDB database that you want to access. Type: - string (or Expression with resultType string). - :type database_name: any - :param username: Username for authentication. Type: string (or Expression with resultType - string). - :type username: any - :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param auth_source: Database to verify the username and password. Type: string (or Expression - with resultType string). - :type auth_source: any - :param port: The TCP port number that the MongoDB server uses to listen for client connections. - The default value is 27017. Type: integer (or Expression with resultType integer), minimum: 0. - :type port: any - :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The - default value is false. Type: boolean (or Expression with resultType boolean). - :type enable_ssl: any - :param allow_self_signed_server_cert: Specifies whether to allow self-signed certificates from - the server. The default value is false. Type: boolean (or Expression with resultType boolean). - :type allow_self_signed_server_cert: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'server': {'required': True}, - 'database_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'database_name': {'key': 'typeProperties.databaseName', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'auth_source': {'key': 'typeProperties.authSource', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, - 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MongoDbLinkedService, self).__init__(**kwargs) - self.type = 'MongoDb' # type: str - self.server = kwargs['server'] - self.authentication_type = kwargs.get('authentication_type', None) - self.database_name = kwargs['database_name'] - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.auth_source = kwargs.get('auth_source', None) - self.port = kwargs.get('port', None) - self.enable_ssl = kwargs.get('enable_ssl', None) - self.allow_self_signed_server_cert = kwargs.get('allow_self_signed_server_cert', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class MongoDbSource(CopySource): - """A copy activity source for a MongoDB database. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: Database query. Should be a SQL-92 query expression. Type: string (or Expression - with resultType string). - :type query: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MongoDbSource, self).__init__(**kwargs) - self.type = 'MongoDbSource' # type: str - self.query = kwargs.get('query', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class MongoDbV2CollectionDataset(Dataset): - """The MongoDB database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param collection: Required. The collection name of the MongoDB database. Type: string (or - Expression with resultType string). - :type collection: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'collection': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'collection': {'key': 'typeProperties.collection', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MongoDbV2CollectionDataset, self).__init__(**kwargs) - self.type = 'MongoDbV2Collection' # type: str - self.collection = kwargs['collection'] - - -class MongoDbV2LinkedService(LinkedService): - """Linked service for MongoDB data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The MongoDB connection string. Type: string, SecureString - or AzureKeyVaultSecretReference. Type: string, SecureString or AzureKeyVaultSecretReference. - :type connection_string: any - :param database: Required. The name of the MongoDB database that you want to access. Type: - string (or Expression with resultType string). - :type database: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - 'database': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MongoDbV2LinkedService, self).__init__(**kwargs) - self.type = 'MongoDbV2' # type: str - self.connection_string = kwargs['connection_string'] - self.database = kwargs['database'] - - -class MongoDbV2Source(CopySource): - """A copy activity source for a MongoDB database. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param filter: Specifies selection filter using query operators. To return all documents in a - collection, omit this parameter or pass an empty document ({}). Type: string (or Expression - with resultType string). - :type filter: any - :param cursor_methods: Cursor methods for Mongodb query. - :type cursor_methods: - ~azure.synapse.artifacts.v2020_12_01.models.MongoDbCursorMethodsProperties - :param batch_size: Specifies the number of documents to return in each batch of the response - from MongoDB instance. In most cases, modifying the batch size will not affect the user or the - application. This property's main purpose is to avoid hit the limitation of response size. - Type: integer (or Expression with resultType integer). - :type batch_size: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'filter': {'key': 'filter', 'type': 'object'}, - 'cursor_methods': {'key': 'cursorMethods', 'type': 'MongoDbCursorMethodsProperties'}, - 'batch_size': {'key': 'batchSize', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MongoDbV2Source, self).__init__(**kwargs) - self.type = 'MongoDbV2Source' # type: str - self.filter = kwargs.get('filter', None) - self.cursor_methods = kwargs.get('cursor_methods', None) - self.batch_size = kwargs.get('batch_size', None) - self.query_timeout = kwargs.get('query_timeout', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class MySqlLinkedService(LinkedService): - """Linked service for MySQL data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The connection string. - :type connection_string: any - :param password: The Azure key vault secret reference of password in connection string. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MySqlLinkedService, self).__init__(**kwargs) - self.type = 'MySql' # type: str - self.connection_string = kwargs['connection_string'] - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class MySqlSource(TabularSource): - """A copy activity source for MySQL databases. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MySqlSource, self).__init__(**kwargs) - self.type = 'MySqlSource' # type: str - self.query = kwargs.get('query', None) - - -class MySqlTableDataset(Dataset): - """The MySQL table dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The MySQL table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(MySqlTableDataset, self).__init__(**kwargs) - self.type = 'MySqlTable' # type: str - self.table_name = kwargs.get('table_name', None) - - -class NetezzaLinkedService(LinkedService): - """Netezza linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: An ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(NetezzaLinkedService, self).__init__(**kwargs) - self.type = 'Netezza' # type: str - self.connection_string = kwargs.get('connection_string', None) - self.pwd = kwargs.get('pwd', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class NetezzaPartitionSettings(msrest.serialization.Model): - """The settings that will be leveraged for Netezza source partitioning. - - :param partition_column_name: The name of the column in integer type that will be used for - proceeding range partitioning. Type: string (or Expression with resultType string). - :type partition_column_name: any - :param partition_upper_bound: The maximum value of column specified in partitionColumnName that - will be used for proceeding range partitioning. Type: string (or Expression with resultType - string). - :type partition_upper_bound: any - :param partition_lower_bound: The minimum value of column specified in partitionColumnName that - will be used for proceeding range partitioning. Type: string (or Expression with resultType - string). - :type partition_lower_bound: any - """ - - _attribute_map = { - 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, - 'partition_upper_bound': {'key': 'partitionUpperBound', 'type': 'object'}, - 'partition_lower_bound': {'key': 'partitionLowerBound', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(NetezzaPartitionSettings, self).__init__(**kwargs) - self.partition_column_name = kwargs.get('partition_column_name', None) - self.partition_upper_bound = kwargs.get('partition_upper_bound', None) - self.partition_lower_bound = kwargs.get('partition_lower_bound', None) - - -class NetezzaSource(TabularSource): - """A copy activity Netezza source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - :param partition_option: The partition mechanism that will be used for Netezza read in - parallel. Possible values include: "None", "DataSlice", "DynamicRange". - :type partition_option: str or - ~azure.synapse.artifacts.v2020_12_01.models.NetezzaPartitionOption - :param partition_settings: The settings that will be leveraged for Netezza source partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.NetezzaPartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'NetezzaPartitionSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(NetezzaSource, self).__init__(**kwargs) - self.type = 'NetezzaSource' # type: str - self.query = kwargs.get('query', None) - self.partition_option = kwargs.get('partition_option', None) - self.partition_settings = kwargs.get('partition_settings', None) - - -class NetezzaTableDataset(Dataset): - """Netezza dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The table name of the Netezza. Type: string (or Expression with resultType - string). - :type table: any - :param schema_type_properties_schema: The schema name of the Netezza. Type: string (or - Expression with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(NetezzaTableDataset, self).__init__(**kwargs) - self.type = 'NetezzaTable' # type: str - self.table_name = kwargs.get('table_name', None) - self.table = kwargs.get('table', None) - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - - -class Notebook(msrest.serialization.Model): - """Notebook. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param description: The description of the notebook. - :type description: str - :param big_data_pool: Big data pool reference. - :type big_data_pool: ~azure.synapse.artifacts.v2020_12_01.models.BigDataPoolReference - :param session_properties: Session properties. - :type session_properties: ~azure.synapse.artifacts.v2020_12_01.models.NotebookSessionProperties - :param metadata: Required. Notebook root-level metadata. - :type metadata: ~azure.synapse.artifacts.v2020_12_01.models.NotebookMetadata - :param nbformat: Required. Notebook format (major number). Incremented between backwards - incompatible changes to the notebook format. - :type nbformat: int - :param nbformat_minor: Required. Notebook format (minor number). Incremented for backward - compatible changes to the notebook format. - :type nbformat_minor: int - :param cells: Required. Array of cells of the current notebook. - :type cells: list[~azure.synapse.artifacts.v2020_12_01.models.NotebookCell] - """ - - _validation = { - 'metadata': {'required': True}, - 'nbformat': {'required': True}, - 'nbformat_minor': {'required': True}, - 'cells': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'description', 'type': 'str'}, - 'big_data_pool': {'key': 'bigDataPool', 'type': 'BigDataPoolReference'}, - 'session_properties': {'key': 'sessionProperties', 'type': 'NotebookSessionProperties'}, - 'metadata': {'key': 'metadata', 'type': 'NotebookMetadata'}, - 'nbformat': {'key': 'nbformat', 'type': 'int'}, - 'nbformat_minor': {'key': 'nbformat_minor', 'type': 'int'}, - 'cells': {'key': 'cells', 'type': '[NotebookCell]'}, - } - - def __init__( - self, - **kwargs - ): - super(Notebook, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.description = kwargs.get('description', None) - self.big_data_pool = kwargs.get('big_data_pool', None) - self.session_properties = kwargs.get('session_properties', None) - self.metadata = kwargs['metadata'] - self.nbformat = kwargs['nbformat'] - self.nbformat_minor = kwargs['nbformat_minor'] - self.cells = kwargs['cells'] - - -class NotebookCell(msrest.serialization.Model): - """Notebook cell. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param cell_type: Required. String identifying the type of cell. - :type cell_type: str - :param metadata: Required. Cell-level metadata. - :type metadata: any - :param source: Required. Contents of the cell, represented as an array of lines. - :type source: list[str] - :param attachments: Attachments associated with the cell. - :type attachments: any - :param outputs: Cell-level output items. - :type outputs: list[~azure.synapse.artifacts.v2020_12_01.models.NotebookCellOutputItem] - """ - - _validation = { - 'cell_type': {'required': True}, - 'metadata': {'required': True}, - 'source': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'cell_type': {'key': 'cell_type', 'type': 'str'}, - 'metadata': {'key': 'metadata', 'type': 'object'}, - 'source': {'key': 'source', 'type': '[str]'}, - 'attachments': {'key': 'attachments', 'type': 'object'}, - 'outputs': {'key': 'outputs', 'type': '[NotebookCellOutputItem]'}, - } - - def __init__( - self, - **kwargs - ): - super(NotebookCell, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.cell_type = kwargs['cell_type'] - self.metadata = kwargs['metadata'] - self.source = kwargs['source'] - self.attachments = kwargs.get('attachments', None) - self.outputs = kwargs.get('outputs', None) - - -class NotebookCellOutputItem(msrest.serialization.Model): - """An item of the notebook cell execution output. - - All required parameters must be populated in order to send to Azure. - - :param name: For output_type=stream, determines the name of stream (stdout / stderr). - :type name: str - :param execution_count: Execution sequence number. - :type execution_count: int - :param output_type: Required. Execution, display, or stream outputs. Possible values include: - "execute_result", "display_data", "stream", "error". - :type output_type: str or ~azure.synapse.artifacts.v2020_12_01.models.CellOutputType - :param text: For output_type=stream, the stream's text output, represented as a string or an - array of strings. - :type text: any - :param data: Output data. Use MIME type as key, and content as value. - :type data: any - :param metadata: Metadata for the output item. - :type metadata: any - """ - - _validation = { - 'output_type': {'required': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'execution_count': {'key': 'execution_count', 'type': 'int'}, - 'output_type': {'key': 'output_type', 'type': 'str'}, - 'text': {'key': 'text', 'type': 'object'}, - 'data': {'key': 'data', 'type': 'object'}, - 'metadata': {'key': 'metadata', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(NotebookCellOutputItem, self).__init__(**kwargs) - self.name = kwargs.get('name', None) - self.execution_count = kwargs.get('execution_count', None) - self.output_type = kwargs['output_type'] - self.text = kwargs.get('text', None) - self.data = kwargs.get('data', None) - self.metadata = kwargs.get('metadata', None) - - -class NotebookKernelSpec(msrest.serialization.Model): - """Kernel information. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Name of the kernel specification. - :type name: str - :param display_name: Required. Name to display in UI. - :type display_name: str - """ - - _validation = { - 'name': {'required': True}, - 'display_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'display_name': {'key': 'display_name', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(NotebookKernelSpec, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.name = kwargs['name'] - self.display_name = kwargs['display_name'] - - -class NotebookLanguageInfo(msrest.serialization.Model): - """Language info. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. The programming language which this kernel runs. - :type name: str - :param codemirror_mode: The codemirror mode to use for code in this language. - :type codemirror_mode: str - """ - - _validation = { - 'name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'codemirror_mode': {'key': 'codemirror_mode', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(NotebookLanguageInfo, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.name = kwargs['name'] - self.codemirror_mode = kwargs.get('codemirror_mode', None) - - -class NotebookListResponse(msrest.serialization.Model): - """A list of Notebook resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of Notebooks. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.NotebookResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[NotebookResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(NotebookListResponse, self).__init__(**kwargs) - self.value = kwargs['value'] - self.next_link = kwargs.get('next_link', None) - - -class NotebookMetadata(msrest.serialization.Model): - """Notebook root-level metadata. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param kernelspec: Kernel information. - :type kernelspec: ~azure.synapse.artifacts.v2020_12_01.models.NotebookKernelSpec - :param language_info: Language info. - :type language_info: ~azure.synapse.artifacts.v2020_12_01.models.NotebookLanguageInfo - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'kernelspec': {'key': 'kernelspec', 'type': 'NotebookKernelSpec'}, - 'language_info': {'key': 'language_info', 'type': 'NotebookLanguageInfo'}, - } - - def __init__( - self, - **kwargs - ): - super(NotebookMetadata, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.kernelspec = kwargs.get('kernelspec', None) - self.language_info = kwargs.get('language_info', None) - - -class NotebookResource(msrest.serialization.Model): - """Notebook resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource Id for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :param name: Required. The name of the resource. - :type name: str - :ivar type: The type of the resource. Ex- Microsoft.Compute/virtualMachines or - Microsoft.Storage/storageAccounts. - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Properties of Notebook. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.Notebook - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'required': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'Notebook'}, - } - - def __init__( - self, - **kwargs - ): - super(NotebookResource, self).__init__(**kwargs) - self.id = None - self.name = kwargs['name'] - self.type = None - self.etag = None - self.properties = kwargs['properties'] - - -class NotebookSessionProperties(msrest.serialization.Model): - """Session properties. - - All required parameters must be populated in order to send to Azure. - - :param driver_memory: Required. Amount of memory to use for the driver process. - :type driver_memory: str - :param driver_cores: Required. Number of cores to use for the driver. - :type driver_cores: int - :param executor_memory: Required. Amount of memory to use per executor process. - :type executor_memory: str - :param executor_cores: Required. Number of cores to use for each executor. - :type executor_cores: int - :param num_executors: Required. Number of executors to launch for this session. - :type num_executors: int - """ - - _validation = { - 'driver_memory': {'required': True}, - 'driver_cores': {'required': True}, - 'executor_memory': {'required': True}, - 'executor_cores': {'required': True}, - 'num_executors': {'required': True}, - } - - _attribute_map = { - 'driver_memory': {'key': 'driverMemory', 'type': 'str'}, - 'driver_cores': {'key': 'driverCores', 'type': 'int'}, - 'executor_memory': {'key': 'executorMemory', 'type': 'str'}, - 'executor_cores': {'key': 'executorCores', 'type': 'int'}, - 'num_executors': {'key': 'numExecutors', 'type': 'int'}, - } - - def __init__( - self, - **kwargs - ): - super(NotebookSessionProperties, self).__init__(**kwargs) - self.driver_memory = kwargs['driver_memory'] - self.driver_cores = kwargs['driver_cores'] - self.executor_memory = kwargs['executor_memory'] - self.executor_cores = kwargs['executor_cores'] - self.num_executors = kwargs['num_executors'] - - -class ODataLinkedService(LinkedService): - """Open Data Protocol (OData) linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param url: Required. The URL of the OData service endpoint. Type: string (or Expression with - resultType string). - :type url: any - :param authentication_type: Type of authentication used to connect to the OData service. - Possible values include: "Basic", "Anonymous", "Windows", "AadServicePrincipal", - "ManagedServiceIdentity". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.ODataAuthenticationType - :param user_name: User name of the OData service. Type: string (or Expression with resultType - string). - :type user_name: any - :param password: Password of the OData service. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: Specify the tenant information (domain name or tenant ID) under which your - application resides. Type: string (or Expression with resultType string). - :type tenant: any - :param service_principal_id: Specify the application id of your application registered in Azure - Active Directory. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param azure_cloud_type: Indicates the azure cloud type of the service principle auth. Allowed - values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data - factory regions’ cloud type. Type: string (or Expression with resultType string). - :type azure_cloud_type: any - :param aad_resource_id: Specify the resource you are requesting authorization to use Directory. - Type: string (or Expression with resultType string). - :type aad_resource_id: any - :param aad_service_principal_credential_type: Specify the credential type (key or cert) is used - for service principal. Possible values include: "ServicePrincipalKey", "ServicePrincipalCert". - :type aad_service_principal_credential_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.ODataAadServicePrincipalCredentialType - :param service_principal_key: Specify the secret of your application registered in Azure Active - Directory. Type: string (or Expression with resultType string). - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param service_principal_embedded_cert: Specify the base64 encoded certificate of your - application registered in Azure Active Directory. Type: string (or Expression with resultType - string). - :type service_principal_embedded_cert: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param service_principal_embedded_cert_password: Specify the password of your certificate if - your certificate has a password and you are using AadServicePrincipal authentication. Type: - string (or Expression with resultType string). - :type service_principal_embedded_cert_password: - ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'azure_cloud_type': {'key': 'typeProperties.azureCloudType', 'type': 'object'}, - 'aad_resource_id': {'key': 'typeProperties.aadResourceId', 'type': 'object'}, - 'aad_service_principal_credential_type': {'key': 'typeProperties.aadServicePrincipalCredentialType', 'type': 'str'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'service_principal_embedded_cert': {'key': 'typeProperties.servicePrincipalEmbeddedCert', 'type': 'SecretBase'}, - 'service_principal_embedded_cert_password': {'key': 'typeProperties.servicePrincipalEmbeddedCertPassword', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ODataLinkedService, self).__init__(**kwargs) - self.type = 'OData' # type: str - self.url = kwargs['url'] - self.authentication_type = kwargs.get('authentication_type', None) - self.user_name = kwargs.get('user_name', None) - self.password = kwargs.get('password', None) - self.tenant = kwargs.get('tenant', None) - self.service_principal_id = kwargs.get('service_principal_id', None) - self.azure_cloud_type = kwargs.get('azure_cloud_type', None) - self.aad_resource_id = kwargs.get('aad_resource_id', None) - self.aad_service_principal_credential_type = kwargs.get('aad_service_principal_credential_type', None) - self.service_principal_key = kwargs.get('service_principal_key', None) - self.service_principal_embedded_cert = kwargs.get('service_principal_embedded_cert', None) - self.service_principal_embedded_cert_password = kwargs.get('service_principal_embedded_cert_password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class ODataResourceDataset(Dataset): - """The Open Data Protocol (OData) resource dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param path: The OData resource path. Type: string (or Expression with resultType string). - :type path: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'path': {'key': 'typeProperties.path', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ODataResourceDataset, self).__init__(**kwargs) - self.type = 'ODataResource' # type: str - self.path = kwargs.get('path', None) - - -class ODataSource(CopySource): - """A copy activity source for OData source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: OData query. For example, "$top=1". Type: string (or Expression with resultType - string). - :type query: any - :param http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout - to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ODataSource, self).__init__(**kwargs) - self.type = 'ODataSource' # type: str - self.query = kwargs.get('query', None) - self.http_request_timeout = kwargs.get('http_request_timeout', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class OdbcLinkedService(LinkedService): - """Open Database Connectivity (ODBC) linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The non-access credential portion of the connection string - as well as an optional encrypted credential. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param authentication_type: Type of authentication used to connect to the ODBC data store. - Possible values are: Anonymous and Basic. Type: string (or Expression with resultType string). - :type authentication_type: any - :param credential: The access credential portion of the connection string specified in - driver-specific property-value format. - :type credential: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param user_name: User name for Basic authentication. Type: string (or Expression with - resultType string). - :type user_name: any - :param password: Password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, - 'credential': {'key': 'typeProperties.credential', 'type': 'SecretBase'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(OdbcLinkedService, self).__init__(**kwargs) - self.type = 'Odbc' # type: str - self.connection_string = kwargs['connection_string'] - self.authentication_type = kwargs.get('authentication_type', None) - self.credential = kwargs.get('credential', None) - self.user_name = kwargs.get('user_name', None) - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class OdbcSink(CopySink): - """A copy activity ODBC sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param pre_copy_script: A query to execute before starting the copy. Type: string (or - Expression with resultType string). - :type pre_copy_script: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(OdbcSink, self).__init__(**kwargs) - self.type = 'OdbcSink' # type: str - self.pre_copy_script = kwargs.get('pre_copy_script', None) - - -class OdbcSource(TabularSource): - """A copy activity source for ODBC databases. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(OdbcSource, self).__init__(**kwargs) - self.type = 'OdbcSource' # type: str - self.query = kwargs.get('query', None) - - -class OdbcTableDataset(Dataset): - """The ODBC table dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The ODBC table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(OdbcTableDataset, self).__init__(**kwargs) - self.type = 'OdbcTable' # type: str - self.table_name = kwargs.get('table_name', None) - - -class Office365Dataset(Dataset): - """The Office365 account. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: Required. Name of the dataset to extract from Office 365. Type: string (or - Expression with resultType string). - :type table_name: any - :param predicate: A predicate expression that can be used to filter the specific rows to - extract from Office 365. Type: string (or Expression with resultType string). - :type predicate: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'table_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'predicate': {'key': 'typeProperties.predicate', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(Office365Dataset, self).__init__(**kwargs) - self.type = 'Office365Table' # type: str - self.table_name = kwargs['table_name'] - self.predicate = kwargs.get('predicate', None) - - -class Office365LinkedService(LinkedService): - """Office365 linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param office365_tenant_id: Required. Azure tenant ID to which the Office 365 account belongs. - Type: string (or Expression with resultType string). - :type office365_tenant_id: any - :param service_principal_tenant_id: Required. Specify the tenant information under which your - Azure AD web application resides. Type: string (or Expression with resultType string). - :type service_principal_tenant_id: any - :param service_principal_id: Required. Specify the application's client ID. Type: string (or - Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: Required. Specify the application's key. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'office365_tenant_id': {'required': True}, - 'service_principal_tenant_id': {'required': True}, - 'service_principal_id': {'required': True}, - 'service_principal_key': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'office365_tenant_id': {'key': 'typeProperties.office365TenantId', 'type': 'object'}, - 'service_principal_tenant_id': {'key': 'typeProperties.servicePrincipalTenantId', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(Office365LinkedService, self).__init__(**kwargs) - self.type = 'Office365' # type: str - self.office365_tenant_id = kwargs['office365_tenant_id'] - self.service_principal_tenant_id = kwargs['service_principal_tenant_id'] - self.service_principal_id = kwargs['service_principal_id'] - self.service_principal_key = kwargs['service_principal_key'] - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class Office365Source(CopySource): - """A copy activity source for an Office 365 service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param allowed_groups: The groups containing all the users. Type: array of strings (or - Expression with resultType array of strings). - :type allowed_groups: any - :param user_scope_filter_uri: The user scope uri. Type: string (or Expression with resultType - string). - :type user_scope_filter_uri: any - :param date_filter_column: The Column to apply the :code:`` and - :code:``. Type: string (or Expression with resultType string). - :type date_filter_column: any - :param start_time: Start time of the requested range for this dataset. Type: string (or - Expression with resultType string). - :type start_time: any - :param end_time: End time of the requested range for this dataset. Type: string (or Expression - with resultType string). - :type end_time: any - :param output_columns: The columns to be read out from the Office 365 table. Type: array of - objects (or Expression with resultType array of objects). Example: [ { "name": "Id" }, { - "name": "CreatedDateTime" } ]. - :type output_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'allowed_groups': {'key': 'allowedGroups', 'type': 'object'}, - 'user_scope_filter_uri': {'key': 'userScopeFilterUri', 'type': 'object'}, - 'date_filter_column': {'key': 'dateFilterColumn', 'type': 'object'}, - 'start_time': {'key': 'startTime', 'type': 'object'}, - 'end_time': {'key': 'endTime', 'type': 'object'}, - 'output_columns': {'key': 'outputColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(Office365Source, self).__init__(**kwargs) - self.type = 'Office365Source' # type: str - self.allowed_groups = kwargs.get('allowed_groups', None) - self.user_scope_filter_uri = kwargs.get('user_scope_filter_uri', None) - self.date_filter_column = kwargs.get('date_filter_column', None) - self.start_time = kwargs.get('start_time', None) - self.end_time = kwargs.get('end_time', None) - self.output_columns = kwargs.get('output_columns', None) - - -class OperationResult(msrest.serialization.Model): - """Operation status for the operation. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar status: Operation status. - :vartype status: str - :param code: Error code. - :type code: str - :param message: Error message. - :type message: str - :param target: Property name/path in request associated with error. - :type target: str - :param details: Array with additional error details. - :type details: list[~azure.synapse.artifacts.v2020_12_01.models.CloudError] - """ - - _validation = { - 'status': {'readonly': True}, - } - - _attribute_map = { - 'status': {'key': 'status', 'type': 'str'}, - 'code': {'key': 'error.code', 'type': 'str'}, - 'message': {'key': 'error.message', 'type': 'str'}, - 'target': {'key': 'error.target', 'type': 'str'}, - 'details': {'key': 'error.details', 'type': '[CloudError]'}, - } - - def __init__( - self, - **kwargs - ): - super(OperationResult, self).__init__(**kwargs) - self.status = None - self.code = kwargs.get('code', None) - self.message = kwargs.get('message', None) - self.target = kwargs.get('target', None) - self.details = kwargs.get('details', None) - - -class OracleLinkedService(LinkedService): - """Oracle database. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param password: The Azure key vault secret reference of password in connection string. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(OracleLinkedService, self).__init__(**kwargs) - self.type = 'Oracle' # type: str - self.connection_string = kwargs['connection_string'] - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class OraclePartitionSettings(msrest.serialization.Model): - """The settings that will be leveraged for Oracle source partitioning. - - :param partition_names: Names of the physical partitions of Oracle table. - :type partition_names: any - :param partition_column_name: The name of the column in integer type that will be used for - proceeding range partitioning. Type: string (or Expression with resultType string). - :type partition_column_name: any - :param partition_upper_bound: The maximum value of column specified in partitionColumnName that - will be used for proceeding range partitioning. Type: string (or Expression with resultType - string). - :type partition_upper_bound: any - :param partition_lower_bound: The minimum value of column specified in partitionColumnName that - will be used for proceeding range partitioning. Type: string (or Expression with resultType - string). - :type partition_lower_bound: any - """ - - _attribute_map = { - 'partition_names': {'key': 'partitionNames', 'type': 'object'}, - 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, - 'partition_upper_bound': {'key': 'partitionUpperBound', 'type': 'object'}, - 'partition_lower_bound': {'key': 'partitionLowerBound', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(OraclePartitionSettings, self).__init__(**kwargs) - self.partition_names = kwargs.get('partition_names', None) - self.partition_column_name = kwargs.get('partition_column_name', None) - self.partition_upper_bound = kwargs.get('partition_upper_bound', None) - self.partition_lower_bound = kwargs.get('partition_lower_bound', None) - - -class OracleServiceCloudLinkedService(LinkedService): - """Oracle Service Cloud linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The URL of the Oracle Service Cloud instance. - :type host: any - :param username: Required. The user name that you use to access Oracle Service Cloud server. - :type username: any - :param password: Required. The password corresponding to the user name that you provided in the - username key. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. Type: boolean (or Expression with resultType boolean). - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. Type: boolean (or Expression with resultType boolean). - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. Type: boolean (or Expression with resultType - boolean). - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'username': {'required': True}, - 'password': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(OracleServiceCloudLinkedService, self).__init__(**kwargs) - self.type = 'OracleServiceCloud' # type: str - self.host = kwargs['host'] - self.username = kwargs['username'] - self.password = kwargs['password'] - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.use_host_verification = kwargs.get('use_host_verification', None) - self.use_peer_verification = kwargs.get('use_peer_verification', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class OracleServiceCloudObjectDataset(Dataset): - """Oracle Service Cloud dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(OracleServiceCloudObjectDataset, self).__init__(**kwargs) - self.type = 'OracleServiceCloudObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class OracleServiceCloudSource(TabularSource): - """A copy activity Oracle Service Cloud source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(OracleServiceCloudSource, self).__init__(**kwargs) - self.type = 'OracleServiceCloudSource' # type: str - self.query = kwargs.get('query', None) - - -class OracleSink(CopySink): - """A copy activity Oracle sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType - string). - :type pre_copy_script: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(OracleSink, self).__init__(**kwargs) - self.type = 'OracleSink' # type: str - self.pre_copy_script = kwargs.get('pre_copy_script', None) - - -class OracleSource(CopySource): - """A copy activity Oracle source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param oracle_reader_query: Oracle reader query. Type: string (or Expression with resultType - string). - :type oracle_reader_query: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param partition_option: The partition mechanism that will be used for Oracle read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". - :type partition_option: str or - ~azure.synapse.artifacts.v2020_12_01.models.OraclePartitionOption - :param partition_settings: The settings that will be leveraged for Oracle source partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.OraclePartitionSettings - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'oracle_reader_query': {'key': 'oracleReaderQuery', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'OraclePartitionSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(OracleSource, self).__init__(**kwargs) - self.type = 'OracleSource' # type: str - self.oracle_reader_query = kwargs.get('oracle_reader_query', None) - self.query_timeout = kwargs.get('query_timeout', None) - self.partition_option = kwargs.get('partition_option', None) - self.partition_settings = kwargs.get('partition_settings', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class OracleTableDataset(Dataset): - """The on-premises Oracle database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param schema_type_properties_schema: The schema name of the on-premises Oracle database. Type: - string (or Expression with resultType string). - :type schema_type_properties_schema: any - :param table: The table name of the on-premises Oracle database. Type: string (or Expression - with resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(OracleTableDataset, self).__init__(**kwargs) - self.type = 'OracleTable' # type: str - self.table_name = kwargs.get('table_name', None) - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - self.table = kwargs.get('table', None) - - -class OrcDataset(Dataset): - """ORC dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param location: The location of the ORC data storage. - :type location: ~azure.synapse.artifacts.v2020_12_01.models.DatasetLocation - :param orc_compression_codec: Possible values include: "none", "zlib", "snappy", "lzo". - :type orc_compression_codec: str or - ~azure.synapse.artifacts.v2020_12_01.models.OrcCompressionCodec - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'location': {'key': 'typeProperties.location', 'type': 'DatasetLocation'}, - 'orc_compression_codec': {'key': 'typeProperties.orcCompressionCodec', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(OrcDataset, self).__init__(**kwargs) - self.type = 'Orc' # type: str - self.location = kwargs.get('location', None) - self.orc_compression_codec = kwargs.get('orc_compression_codec', None) - - -class OrcFormat(DatasetStorageFormat): - """The data stored in Optimized Row Columnar (ORC) format. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage format.Constant filled by server. - :type type: str - :param serializer: Serializer. Type: string (or Expression with resultType string). - :type serializer: any - :param deserializer: Deserializer. Type: string (or Expression with resultType string). - :type deserializer: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'serializer': {'key': 'serializer', 'type': 'object'}, - 'deserializer': {'key': 'deserializer', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(OrcFormat, self).__init__(**kwargs) - self.type = 'OrcFormat' # type: str - - -class OrcSink(CopySink): - """A copy activity ORC sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: ORC store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreWriteSettings - :param format_settings: ORC format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.OrcWriteSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreWriteSettings'}, - 'format_settings': {'key': 'formatSettings', 'type': 'OrcWriteSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(OrcSink, self).__init__(**kwargs) - self.type = 'OrcSink' # type: str - self.store_settings = kwargs.get('store_settings', None) - self.format_settings = kwargs.get('format_settings', None) - - -class OrcSource(CopySource): - """A copy activity ORC source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: ORC store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(OrcSource, self).__init__(**kwargs) - self.type = 'OrcSource' # type: str - self.store_settings = kwargs.get('store_settings', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class OrcWriteSettings(FormatWriteSettings): - """Orc write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param max_rows_per_file: Limit the written file's row count to be smaller than or equal to the - specified count. Type: integer (or Expression with resultType integer). - :type max_rows_per_file: any - :param file_name_prefix: Specifies the file name pattern - :code:``_:code:``.:code:`` when copy from non-file - based store without partitionOptions. Type: string (or Expression with resultType string). - :type file_name_prefix: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_rows_per_file': {'key': 'maxRowsPerFile', 'type': 'object'}, - 'file_name_prefix': {'key': 'fileNamePrefix', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(OrcWriteSettings, self).__init__(**kwargs) - self.type = 'OrcWriteSettings' # type: str - self.max_rows_per_file = kwargs.get('max_rows_per_file', None) - self.file_name_prefix = kwargs.get('file_name_prefix', None) - - -class ParameterSpecification(msrest.serialization.Model): - """Definition of a single parameter for an entity. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Parameter type. Possible values include: "Object", "String", "Int", - "Float", "Bool", "Array", "SecureString". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.ParameterType - :param default_value: Default value of parameter. - :type default_value: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'default_value': {'key': 'defaultValue', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ParameterSpecification, self).__init__(**kwargs) - self.type = kwargs['type'] - self.default_value = kwargs.get('default_value', None) - - -class ParquetDataset(Dataset): - """Parquet dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param location: The location of the parquet storage. - :type location: ~azure.synapse.artifacts.v2020_12_01.models.DatasetLocation - :param compression_codec: A string from ParquetCompressionCodecEnum or an expression. - :type compression_codec: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'location': {'key': 'typeProperties.location', 'type': 'DatasetLocation'}, - 'compression_codec': {'key': 'typeProperties.compressionCodec', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ParquetDataset, self).__init__(**kwargs) - self.type = 'Parquet' # type: str - self.location = kwargs.get('location', None) - self.compression_codec = kwargs.get('compression_codec', None) - - -class ParquetFormat(DatasetStorageFormat): - """The data stored in Parquet format. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage format.Constant filled by server. - :type type: str - :param serializer: Serializer. Type: string (or Expression with resultType string). - :type serializer: any - :param deserializer: Deserializer. Type: string (or Expression with resultType string). - :type deserializer: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'serializer': {'key': 'serializer', 'type': 'object'}, - 'deserializer': {'key': 'deserializer', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ParquetFormat, self).__init__(**kwargs) - self.type = 'ParquetFormat' # type: str - - -class ParquetSink(CopySink): - """A copy activity Parquet sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Parquet store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreWriteSettings - :param format_settings: Parquet format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.ParquetWriteSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreWriteSettings'}, - 'format_settings': {'key': 'formatSettings', 'type': 'ParquetWriteSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(ParquetSink, self).__init__(**kwargs) - self.type = 'ParquetSink' # type: str - self.store_settings = kwargs.get('store_settings', None) - self.format_settings = kwargs.get('format_settings', None) - - -class ParquetSource(CopySource): - """A copy activity Parquet source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Parquet store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ParquetSource, self).__init__(**kwargs) - self.type = 'ParquetSource' # type: str - self.store_settings = kwargs.get('store_settings', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class ParquetWriteSettings(FormatWriteSettings): - """Parquet write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param max_rows_per_file: Limit the written file's row count to be smaller than or equal to the - specified count. Type: integer (or Expression with resultType integer). - :type max_rows_per_file: any - :param file_name_prefix: Specifies the file name pattern - :code:``_:code:``.:code:`` when copy from non-file - based store without partitionOptions. Type: string (or Expression with resultType string). - :type file_name_prefix: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_rows_per_file': {'key': 'maxRowsPerFile', 'type': 'object'}, - 'file_name_prefix': {'key': 'fileNamePrefix', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ParquetWriteSettings, self).__init__(**kwargs) - self.type = 'ParquetWriteSettings' # type: str - self.max_rows_per_file = kwargs.get('max_rows_per_file', None) - self.file_name_prefix = kwargs.get('file_name_prefix', None) - - -class PaypalLinkedService(LinkedService): - """Paypal Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The URL of the PayPal instance. (i.e. api.sandbox.paypal.com). - :type host: any - :param client_id: Required. The client ID associated with your PayPal application. - :type client_id: any - :param client_secret: The client secret associated with your PayPal application. - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'client_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(PaypalLinkedService, self).__init__(**kwargs) - self.type = 'Paypal' # type: str - self.host = kwargs['host'] - self.client_id = kwargs['client_id'] - self.client_secret = kwargs.get('client_secret', None) - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.use_host_verification = kwargs.get('use_host_verification', None) - self.use_peer_verification = kwargs.get('use_peer_verification', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class PaypalObjectDataset(Dataset): - """Paypal Service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(PaypalObjectDataset, self).__init__(**kwargs) - self.type = 'PaypalObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class PaypalSource(TabularSource): - """A copy activity Paypal Service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(PaypalSource, self).__init__(**kwargs) - self.type = 'PaypalSource' # type: str - self.query = kwargs.get('query', None) - - -class PhoenixLinkedService(LinkedService): - """Phoenix server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The IP address or host name of the Phoenix server. (i.e. - 192.168.222.160). - :type host: any - :param port: The TCP port that the Phoenix server uses to listen for client connections. The - default value is 8765. - :type port: any - :param http_path: The partial URL corresponding to the Phoenix server. (i.e. - /gateway/sandbox/phoenix/version). The default value is hbasephoenix if using - WindowsAzureHDInsightService. - :type http_path: any - :param authentication_type: Required. The authentication mechanism used to connect to the - Phoenix server. Possible values include: "Anonymous", "UsernameAndPassword", - "WindowsAzureHDInsightService". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.PhoenixAuthenticationType - :param username: The user name used to connect to the Phoenix server. - :type username: any - :param password: The password corresponding to the user name. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The - default value is false. - :type enable_ssl: any - :param trusted_cert_path: The full path of the .pem file containing trusted CA certificates for - verifying the server when connecting over SSL. This property can only be set when using SSL on - self-hosted IR. The default value is the cacerts.pem file installed with the IR. - :type trusted_cert_path: any - :param use_system_trust_store: Specifies whether to use a CA certificate from the system trust - store or from a specified PEM file. The default value is false. - :type use_system_trust_store: any - :param allow_host_name_cn_mismatch: Specifies whether to require a CA-issued SSL certificate - name to match the host name of the server when connecting over SSL. The default value is false. - :type allow_host_name_cn_mismatch: any - :param allow_self_signed_server_cert: Specifies whether to allow self-signed certificates from - the server. The default value is false. - :type allow_self_signed_server_cert: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'http_path': {'key': 'typeProperties.httpPath', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, - 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, - 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, - 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, - 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(PhoenixLinkedService, self).__init__(**kwargs) - self.type = 'Phoenix' # type: str - self.host = kwargs['host'] - self.port = kwargs.get('port', None) - self.http_path = kwargs.get('http_path', None) - self.authentication_type = kwargs['authentication_type'] - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.enable_ssl = kwargs.get('enable_ssl', None) - self.trusted_cert_path = kwargs.get('trusted_cert_path', None) - self.use_system_trust_store = kwargs.get('use_system_trust_store', None) - self.allow_host_name_cn_mismatch = kwargs.get('allow_host_name_cn_mismatch', None) - self.allow_self_signed_server_cert = kwargs.get('allow_self_signed_server_cert', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class PhoenixObjectDataset(Dataset): - """Phoenix server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The table name of the Phoenix. Type: string (or Expression with resultType - string). - :type table: any - :param schema_type_properties_schema: The schema name of the Phoenix. Type: string (or - Expression with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(PhoenixObjectDataset, self).__init__(**kwargs) - self.type = 'PhoenixObject' # type: str - self.table_name = kwargs.get('table_name', None) - self.table = kwargs.get('table', None) - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - - -class PhoenixSource(TabularSource): - """A copy activity Phoenix server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(PhoenixSource, self).__init__(**kwargs) - self.type = 'PhoenixSource' # type: str - self.query = kwargs.get('query', None) - - -class PipelineFolder(msrest.serialization.Model): - """The folder that this Pipeline is in. If not specified, Pipeline will appear at the root level. - - :param name: The name of the folder that this Pipeline is in. - :type name: str - """ - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(PipelineFolder, self).__init__(**kwargs) - self.name = kwargs.get('name', None) - - -class PipelineListResponse(msrest.serialization.Model): - """A list of pipeline resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of pipelines. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.PipelineResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[PipelineResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(PipelineListResponse, self).__init__(**kwargs) - self.value = kwargs['value'] - self.next_link = kwargs.get('next_link', None) - - -class PipelineReference(msrest.serialization.Model): - """Pipeline reference type. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Pipeline reference type. Possible values include: "PipelineReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.PipelineReferenceType - :param reference_name: Required. Reference pipeline name. - :type reference_name: str - :param name: Reference name. - :type name: str - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(PipelineReference, self).__init__(**kwargs) - self.type = kwargs['type'] - self.reference_name = kwargs['reference_name'] - self.name = kwargs.get('name', None) - - -class PipelineResource(SubResource): - """Pipeline resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param description: The description of the pipeline. - :type description: str - :param activities: List of activities in pipeline. - :type activities: list[~azure.synapse.artifacts.v2020_12_01.models.Activity] - :param parameters: List of parameters for pipeline. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param variables: List of variables for pipeline. - :type variables: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.VariableSpecification] - :param concurrency: The max number of concurrent runs for the pipeline. - :type concurrency: int - :param annotations: List of tags that can be used for describing the Pipeline. - :type annotations: list[any] - :param run_dimensions: Dimensions emitted by Pipeline. - :type run_dimensions: dict[str, any] - :param folder: The folder that this Pipeline is in. If not specified, Pipeline will appear at - the root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.PipelineFolder - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'concurrency': {'minimum': 1}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'properties.description', 'type': 'str'}, - 'activities': {'key': 'properties.activities', 'type': '[Activity]'}, - 'parameters': {'key': 'properties.parameters', 'type': '{ParameterSpecification}'}, - 'variables': {'key': 'properties.variables', 'type': '{VariableSpecification}'}, - 'concurrency': {'key': 'properties.concurrency', 'type': 'int'}, - 'annotations': {'key': 'properties.annotations', 'type': '[object]'}, - 'run_dimensions': {'key': 'properties.runDimensions', 'type': '{object}'}, - 'folder': {'key': 'properties.folder', 'type': 'PipelineFolder'}, - } - - def __init__( - self, - **kwargs - ): - super(PipelineResource, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.description = kwargs.get('description', None) - self.activities = kwargs.get('activities', None) - self.parameters = kwargs.get('parameters', None) - self.variables = kwargs.get('variables', None) - self.concurrency = kwargs.get('concurrency', None) - self.annotations = kwargs.get('annotations', None) - self.run_dimensions = kwargs.get('run_dimensions', None) - self.folder = kwargs.get('folder', None) - - -class PipelineRun(msrest.serialization.Model): - """Information about a pipeline run. - - Variables are only populated by the server, and will be ignored when sending a request. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :ivar run_id: Identifier of a run. - :vartype run_id: str - :ivar run_group_id: Identifier that correlates all the recovery runs of a pipeline run. - :vartype run_group_id: str - :ivar is_latest: Indicates if the recovered pipeline run is the latest in its group. - :vartype is_latest: bool - :ivar pipeline_name: The pipeline name. - :vartype pipeline_name: str - :ivar parameters: The full or partial list of parameter name, value pair used in the pipeline - run. - :vartype parameters: dict[str, str] - :ivar invoked_by: Entity that started the pipeline run. - :vartype invoked_by: ~azure.synapse.artifacts.v2020_12_01.models.PipelineRunInvokedBy - :ivar last_updated: The last updated timestamp for the pipeline run event in ISO8601 format. - :vartype last_updated: ~datetime.datetime - :ivar run_start: The start time of a pipeline run in ISO8601 format. - :vartype run_start: ~datetime.datetime - :ivar run_end: The end time of a pipeline run in ISO8601 format. - :vartype run_end: ~datetime.datetime - :ivar duration_in_ms: The duration of a pipeline run. - :vartype duration_in_ms: int - :ivar status: The status of a pipeline run. - :vartype status: str - :ivar message: The message from a pipeline run. - :vartype message: str - """ - - _validation = { - 'run_id': {'readonly': True}, - 'run_group_id': {'readonly': True}, - 'is_latest': {'readonly': True}, - 'pipeline_name': {'readonly': True}, - 'parameters': {'readonly': True}, - 'invoked_by': {'readonly': True}, - 'last_updated': {'readonly': True}, - 'run_start': {'readonly': True}, - 'run_end': {'readonly': True}, - 'duration_in_ms': {'readonly': True}, - 'status': {'readonly': True}, - 'message': {'readonly': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'run_id': {'key': 'runId', 'type': 'str'}, - 'run_group_id': {'key': 'runGroupId', 'type': 'str'}, - 'is_latest': {'key': 'isLatest', 'type': 'bool'}, - 'pipeline_name': {'key': 'pipelineName', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{str}'}, - 'invoked_by': {'key': 'invokedBy', 'type': 'PipelineRunInvokedBy'}, - 'last_updated': {'key': 'lastUpdated', 'type': 'iso-8601'}, - 'run_start': {'key': 'runStart', 'type': 'iso-8601'}, - 'run_end': {'key': 'runEnd', 'type': 'iso-8601'}, - 'duration_in_ms': {'key': 'durationInMs', 'type': 'int'}, - 'status': {'key': 'status', 'type': 'str'}, - 'message': {'key': 'message', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(PipelineRun, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.run_id = None - self.run_group_id = None - self.is_latest = None - self.pipeline_name = None - self.parameters = None - self.invoked_by = None - self.last_updated = None - self.run_start = None - self.run_end = None - self.duration_in_ms = None - self.status = None - self.message = None - - -class PipelineRunInvokedBy(msrest.serialization.Model): - """Provides entity name and id that started the pipeline run. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar name: Name of the entity that started the pipeline run. - :vartype name: str - :ivar id: The ID of the entity that started the run. - :vartype id: str - :ivar invoked_by_type: The type of the entity that started the run. - :vartype invoked_by_type: str - """ - - _validation = { - 'name': {'readonly': True}, - 'id': {'readonly': True}, - 'invoked_by_type': {'readonly': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'id': {'key': 'id', 'type': 'str'}, - 'invoked_by_type': {'key': 'invokedByType', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(PipelineRunInvokedBy, self).__init__(**kwargs) - self.name = None - self.id = None - self.invoked_by_type = None - - -class PipelineRunsQueryResponse(msrest.serialization.Model): - """A list pipeline runs. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of pipeline runs. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.PipelineRun] - :param continuation_token: The continuation token for getting the next page of results, if any - remaining results exist, null otherwise. - :type continuation_token: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[PipelineRun]'}, - 'continuation_token': {'key': 'continuationToken', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(PipelineRunsQueryResponse, self).__init__(**kwargs) - self.value = kwargs['value'] - self.continuation_token = kwargs.get('continuation_token', None) - - -class PolybaseSettings(msrest.serialization.Model): - """PolyBase settings. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param reject_type: Reject type. Possible values include: "value", "percentage". - :type reject_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.PolybaseSettingsRejectType - :param reject_value: Specifies the value or the percentage of rows that can be rejected before - the query fails. Type: number (or Expression with resultType number), minimum: 0. - :type reject_value: any - :param reject_sample_value: Determines the number of rows to attempt to retrieve before the - PolyBase recalculates the percentage of rejected rows. Type: integer (or Expression with - resultType integer), minimum: 0. - :type reject_sample_value: any - :param use_type_default: Specifies how to handle missing values in delimited text files when - PolyBase retrieves data from the text file. Type: boolean (or Expression with resultType - boolean). - :type use_type_default: any - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'reject_type': {'key': 'rejectType', 'type': 'str'}, - 'reject_value': {'key': 'rejectValue', 'type': 'object'}, - 'reject_sample_value': {'key': 'rejectSampleValue', 'type': 'object'}, - 'use_type_default': {'key': 'useTypeDefault', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(PolybaseSettings, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.reject_type = kwargs.get('reject_type', None) - self.reject_value = kwargs.get('reject_value', None) - self.reject_sample_value = kwargs.get('reject_sample_value', None) - self.use_type_default = kwargs.get('use_type_default', None) - - -class PostgreSqlLinkedService(LinkedService): - """Linked service for PostgreSQL data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The connection string. - :type connection_string: any - :param password: The Azure key vault secret reference of password in connection string. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(PostgreSqlLinkedService, self).__init__(**kwargs) - self.type = 'PostgreSql' # type: str - self.connection_string = kwargs['connection_string'] - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class PostgreSqlSource(TabularSource): - """A copy activity source for PostgreSQL databases. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(PostgreSqlSource, self).__init__(**kwargs) - self.type = 'PostgreSqlSource' # type: str - self.query = kwargs.get('query', None) - - -class PostgreSqlTableDataset(Dataset): - """The PostgreSQL table dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The PostgreSQL table name. Type: string (or Expression with resultType string). - :type table: any - :param schema_type_properties_schema: The PostgreSQL schema name. Type: string (or Expression - with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(PostgreSqlTableDataset, self).__init__(**kwargs) - self.type = 'PostgreSqlTable' # type: str - self.table_name = kwargs.get('table_name', None) - self.table = kwargs.get('table', None) - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - - -class PrestoLinkedService(LinkedService): - """Presto server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The IP address or host name of the Presto server. (i.e. - 192.168.222.160). - :type host: any - :param server_version: Required. The version of the Presto server. (i.e. 0.148-t). - :type server_version: any - :param catalog: Required. The catalog context for all request against the server. - :type catalog: any - :param port: The TCP port that the Presto server uses to listen for client connections. The - default value is 8080. - :type port: any - :param authentication_type: Required. The authentication mechanism used to connect to the - Presto server. Possible values include: "Anonymous", "LDAP". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.PrestoAuthenticationType - :param username: The user name used to connect to the Presto server. - :type username: any - :param password: The password corresponding to the user name. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The - default value is false. - :type enable_ssl: any - :param trusted_cert_path: The full path of the .pem file containing trusted CA certificates for - verifying the server when connecting over SSL. This property can only be set when using SSL on - self-hosted IR. The default value is the cacerts.pem file installed with the IR. - :type trusted_cert_path: any - :param use_system_trust_store: Specifies whether to use a CA certificate from the system trust - store or from a specified PEM file. The default value is false. - :type use_system_trust_store: any - :param allow_host_name_cn_mismatch: Specifies whether to require a CA-issued SSL certificate - name to match the host name of the server when connecting over SSL. The default value is false. - :type allow_host_name_cn_mismatch: any - :param allow_self_signed_server_cert: Specifies whether to allow self-signed certificates from - the server. The default value is false. - :type allow_self_signed_server_cert: any - :param time_zone_id: The local time zone used by the connection. Valid values for this option - are specified in the IANA Time Zone Database. The default value is the system time zone. - :type time_zone_id: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'server_version': {'required': True}, - 'catalog': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'server_version': {'key': 'typeProperties.serverVersion', 'type': 'object'}, - 'catalog': {'key': 'typeProperties.catalog', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, - 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, - 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, - 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, - 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, - 'time_zone_id': {'key': 'typeProperties.timeZoneID', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(PrestoLinkedService, self).__init__(**kwargs) - self.type = 'Presto' # type: str - self.host = kwargs['host'] - self.server_version = kwargs['server_version'] - self.catalog = kwargs['catalog'] - self.port = kwargs.get('port', None) - self.authentication_type = kwargs['authentication_type'] - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.enable_ssl = kwargs.get('enable_ssl', None) - self.trusted_cert_path = kwargs.get('trusted_cert_path', None) - self.use_system_trust_store = kwargs.get('use_system_trust_store', None) - self.allow_host_name_cn_mismatch = kwargs.get('allow_host_name_cn_mismatch', None) - self.allow_self_signed_server_cert = kwargs.get('allow_self_signed_server_cert', None) - self.time_zone_id = kwargs.get('time_zone_id', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class PrestoObjectDataset(Dataset): - """Presto server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The table name of the Presto. Type: string (or Expression with resultType - string). - :type table: any - :param schema_type_properties_schema: The schema name of the Presto. Type: string (or - Expression with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(PrestoObjectDataset, self).__init__(**kwargs) - self.type = 'PrestoObject' # type: str - self.table_name = kwargs.get('table_name', None) - self.table = kwargs.get('table', None) - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - - -class PrestoSource(TabularSource): - """A copy activity Presto server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(PrestoSource, self).__init__(**kwargs) - self.type = 'PrestoSource' # type: str - self.query = kwargs.get('query', None) - - -class PrivateEndpoint(msrest.serialization.Model): - """Private endpoint details. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar id: Resource id of the private endpoint. - :vartype id: str - """ - - _validation = { - 'id': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(PrivateEndpoint, self).__init__(**kwargs) - self.id = None - - -class ProxyResource(Resource): - """The resource model definition for a Azure Resource Manager proxy resource. It will not have tags and a location. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(ProxyResource, self).__init__(**kwargs) - - -class PrivateEndpointConnection(ProxyResource): - """A private endpoint connection. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :param private_endpoint: The private endpoint which the connection belongs to. - :type private_endpoint: ~azure.synapse.artifacts.v2020_12_01.models.PrivateEndpoint - :param private_link_service_connection_state: Connection state of the private endpoint - connection. - :type private_link_service_connection_state: - ~azure.synapse.artifacts.v2020_12_01.models.PrivateLinkServiceConnectionState - :ivar provisioning_state: Provisioning state of the private endpoint connection. - :vartype provisioning_state: str - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'provisioning_state': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'private_endpoint': {'key': 'properties.privateEndpoint', 'type': 'PrivateEndpoint'}, - 'private_link_service_connection_state': {'key': 'properties.privateLinkServiceConnectionState', 'type': 'PrivateLinkServiceConnectionState'}, - 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(PrivateEndpointConnection, self).__init__(**kwargs) - self.private_endpoint = kwargs.get('private_endpoint', None) - self.private_link_service_connection_state = kwargs.get('private_link_service_connection_state', None) - self.provisioning_state = None - - -class PrivateLinkServiceConnectionState(msrest.serialization.Model): - """Connection state details of the private endpoint. - - Variables are only populated by the server, and will be ignored when sending a request. - - :param status: The private link service connection status. - :type status: str - :param description: The private link service connection description. - :type description: str - :ivar actions_required: The actions required for private link service connection. - :vartype actions_required: str - """ - - _validation = { - 'actions_required': {'readonly': True}, - } - - _attribute_map = { - 'status': {'key': 'status', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'actions_required': {'key': 'actionsRequired', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(PrivateLinkServiceConnectionState, self).__init__(**kwargs) - self.status = kwargs.get('status', None) - self.description = kwargs.get('description', None) - self.actions_required = None - - -class PurviewConfiguration(msrest.serialization.Model): - """Purview Configuration. - - :param purview_resource_id: Purview Resource ID. - :type purview_resource_id: str - """ - - _attribute_map = { - 'purview_resource_id': {'key': 'purviewResourceId', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(PurviewConfiguration, self).__init__(**kwargs) - self.purview_resource_id = kwargs.get('purview_resource_id', None) - - -class QueryDataFlowDebugSessionsResponse(msrest.serialization.Model): - """A list of active debug sessions. - - :param value: Array with all active debug sessions. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.DataFlowDebugSessionInfo] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _attribute_map = { - 'value': {'key': 'value', 'type': '[DataFlowDebugSessionInfo]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(QueryDataFlowDebugSessionsResponse, self).__init__(**kwargs) - self.value = kwargs.get('value', None) - self.next_link = kwargs.get('next_link', None) - - -class QuickBooksLinkedService(LinkedService): - """QuickBooks server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_properties: Properties used to connect to QuickBooks. It is mutually - exclusive with any other properties in the linked service. Type: object. - :type connection_properties: any - :param endpoint: Required. The endpoint of the QuickBooks server. (i.e. - quickbooks.api.intuit.com). - :type endpoint: any - :param company_id: Required. The company ID of the QuickBooks company to authorize. - :type company_id: any - :param consumer_key: Required. The consumer key for OAuth 1.0 authentication. - :type consumer_key: any - :param consumer_secret: Required. The consumer secret for OAuth 1.0 authentication. - :type consumer_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param access_token: Required. The access token for OAuth 1.0 authentication. - :type access_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param access_token_secret: Required. The access token secret for OAuth 1.0 authentication. - :type access_token_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'endpoint': {'required': True}, - 'company_id': {'required': True}, - 'consumer_key': {'required': True}, - 'consumer_secret': {'required': True}, - 'access_token': {'required': True}, - 'access_token_secret': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_properties': {'key': 'typeProperties.connectionProperties', 'type': 'object'}, - 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, - 'company_id': {'key': 'typeProperties.companyId', 'type': 'object'}, - 'consumer_key': {'key': 'typeProperties.consumerKey', 'type': 'object'}, - 'consumer_secret': {'key': 'typeProperties.consumerSecret', 'type': 'SecretBase'}, - 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, - 'access_token_secret': {'key': 'typeProperties.accessTokenSecret', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(QuickBooksLinkedService, self).__init__(**kwargs) - self.type = 'QuickBooks' # type: str - self.connection_properties = kwargs.get('connection_properties', None) - self.endpoint = kwargs['endpoint'] - self.company_id = kwargs['company_id'] - self.consumer_key = kwargs['consumer_key'] - self.consumer_secret = kwargs['consumer_secret'] - self.access_token = kwargs['access_token'] - self.access_token_secret = kwargs['access_token_secret'] - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class QuickBooksObjectDataset(Dataset): - """QuickBooks server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(QuickBooksObjectDataset, self).__init__(**kwargs) - self.type = 'QuickBooksObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class QuickBooksSource(TabularSource): - """A copy activity QuickBooks server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(QuickBooksSource, self).__init__(**kwargs) - self.type = 'QuickBooksSource' # type: str - self.query = kwargs.get('query', None) - - -class RecurrenceSchedule(msrest.serialization.Model): - """The recurrence schedule. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param minutes: The minutes. - :type minutes: list[int] - :param hours: The hours. - :type hours: list[int] - :param week_days: The days of the week. - :type week_days: list[str or ~azure.synapse.artifacts.v2020_12_01.models.DayOfWeek] - :param month_days: The month days. - :type month_days: list[int] - :param monthly_occurrences: The monthly occurrences. - :type monthly_occurrences: - list[~azure.synapse.artifacts.v2020_12_01.models.RecurrenceScheduleOccurrence] - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'minutes': {'key': 'minutes', 'type': '[int]'}, - 'hours': {'key': 'hours', 'type': '[int]'}, - 'week_days': {'key': 'weekDays', 'type': '[str]'}, - 'month_days': {'key': 'monthDays', 'type': '[int]'}, - 'monthly_occurrences': {'key': 'monthlyOccurrences', 'type': '[RecurrenceScheduleOccurrence]'}, - } - - def __init__( - self, - **kwargs - ): - super(RecurrenceSchedule, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.minutes = kwargs.get('minutes', None) - self.hours = kwargs.get('hours', None) - self.week_days = kwargs.get('week_days', None) - self.month_days = kwargs.get('month_days', None) - self.monthly_occurrences = kwargs.get('monthly_occurrences', None) - - -class RecurrenceScheduleOccurrence(msrest.serialization.Model): - """The recurrence schedule occurrence. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param day: The day of the week. Possible values include: "Sunday", "Monday", "Tuesday", - "Wednesday", "Thursday", "Friday", "Saturday". - :type day: str or ~azure.synapse.artifacts.v2020_12_01.models.DayOfWeek - :param occurrence: The occurrence. - :type occurrence: int - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'day': {'key': 'day', 'type': 'str'}, - 'occurrence': {'key': 'occurrence', 'type': 'int'}, - } - - def __init__( - self, - **kwargs - ): - super(RecurrenceScheduleOccurrence, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.day = kwargs.get('day', None) - self.occurrence = kwargs.get('occurrence', None) - - -class RedirectIncompatibleRowSettings(msrest.serialization.Model): - """Redirect incompatible row settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param linked_service_name: Required. Name of the Azure Storage, Storage SAS, or Azure Data - Lake Store linked service used for redirecting incompatible row. Must be specified if - redirectIncompatibleRowSettings is specified. Type: string (or Expression with resultType - string). - :type linked_service_name: any - :param path: The path for storing the redirect incompatible row data. Type: string (or - Expression with resultType string). - :type path: any - """ - - _validation = { - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'object'}, - 'path': {'key': 'path', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(RedirectIncompatibleRowSettings, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.linked_service_name = kwargs['linked_service_name'] - self.path = kwargs.get('path', None) - - -class RedshiftUnloadSettings(msrest.serialization.Model): - """The Amazon S3 settings needed for the interim Amazon S3 when copying from Amazon Redshift with unload. With this, data from Amazon Redshift source will be unloaded into S3 first and then copied into the targeted sink from the interim S3. - - All required parameters must be populated in order to send to Azure. - - :param s3_linked_service_name: Required. The name of the Amazon S3 linked service which will be - used for the unload operation when copying from the Amazon Redshift source. - :type s3_linked_service_name: - ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param bucket_name: Required. The bucket of the interim Amazon S3 which will be used to store - the unloaded data from Amazon Redshift source. The bucket must be in the same region as the - Amazon Redshift source. Type: string (or Expression with resultType string). - :type bucket_name: any - """ - - _validation = { - 's3_linked_service_name': {'required': True}, - 'bucket_name': {'required': True}, - } - - _attribute_map = { - 's3_linked_service_name': {'key': 's3LinkedServiceName', 'type': 'LinkedServiceReference'}, - 'bucket_name': {'key': 'bucketName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(RedshiftUnloadSettings, self).__init__(**kwargs) - self.s3_linked_service_name = kwargs['s3_linked_service_name'] - self.bucket_name = kwargs['bucket_name'] - - -class RelationalSource(CopySource): - """A copy activity source for various relational databases. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(RelationalSource, self).__init__(**kwargs) - self.type = 'RelationalSource' # type: str - self.query = kwargs.get('query', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class RelationalTableDataset(Dataset): - """The relational table dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The relational table name. Type: string (or Expression with resultType - string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(RelationalTableDataset, self).__init__(**kwargs) - self.type = 'RelationalTable' # type: str - self.table_name = kwargs.get('table_name', None) - - -class RerunTriggerListResponse(msrest.serialization.Model): - """A list of rerun triggers. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of rerun triggers. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.RerunTriggerResource] - :ivar next_link: The continuation token for getting the next page of results, if any remaining - results exist, null otherwise. - :vartype next_link: str - """ - - _validation = { - 'value': {'required': True}, - 'next_link': {'readonly': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[RerunTriggerResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(RerunTriggerListResponse, self).__init__(**kwargs) - self.value = kwargs['value'] - self.next_link = None - - -class RerunTriggerResource(SubResource): - """RerunTrigger resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Properties of the rerun trigger. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.RerunTumblingWindowTrigger - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'RerunTumblingWindowTrigger'}, - } - - def __init__( - self, - **kwargs - ): - super(RerunTriggerResource, self).__init__(**kwargs) - self.properties = kwargs['properties'] - - -class RerunTumblingWindowTrigger(Trigger): - """Trigger that schedules pipeline reruns for all fixed time interval windows from a requested start time to requested end time. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Trigger type.Constant filled by server. - :type type: str - :param description: Trigger description. - :type description: str - :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are - called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRuntimeState - :param annotations: List of tags that can be used for describing the trigger. - :type annotations: list[any] - :param parent_trigger: Required. The parent trigger reference. - :type parent_trigger: any - :param requested_start_time: Required. The start time for the time period for which restatement - is initiated. Only UTC time is currently supported. - :type requested_start_time: ~datetime.datetime - :param requested_end_time: Required. The end time for the time period for which restatement is - initiated. Only UTC time is currently supported. - :type requested_end_time: ~datetime.datetime - :param rerun_concurrency: Required. The max number of parallel time windows (ready for - execution) for which a rerun is triggered. - :type rerun_concurrency: int - """ - - _validation = { - 'type': {'required': True}, - 'runtime_state': {'readonly': True}, - 'parent_trigger': {'required': True}, - 'requested_start_time': {'required': True}, - 'requested_end_time': {'required': True}, - 'rerun_concurrency': {'required': True, 'maximum': 50, 'minimum': 1}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'parent_trigger': {'key': 'typeProperties.parentTrigger', 'type': 'object'}, - 'requested_start_time': {'key': 'typeProperties.requestedStartTime', 'type': 'iso-8601'}, - 'requested_end_time': {'key': 'typeProperties.requestedEndTime', 'type': 'iso-8601'}, - 'rerun_concurrency': {'key': 'typeProperties.rerunConcurrency', 'type': 'int'}, - } - - def __init__( - self, - **kwargs - ): - super(RerunTumblingWindowTrigger, self).__init__(**kwargs) - self.type = 'RerunTumblingWindowTrigger' # type: str - self.parent_trigger = kwargs['parent_trigger'] - self.requested_start_time = kwargs['requested_start_time'] - self.requested_end_time = kwargs['requested_end_time'] - self.rerun_concurrency = kwargs['rerun_concurrency'] - - -class RerunTumblingWindowTriggerActionParameters(msrest.serialization.Model): - """Rerun tumbling window trigger Parameters. - - All required parameters must be populated in order to send to Azure. - - :param start_time: Required. The start time for the time period for which restatement is - initiated. Only UTC time is currently supported. - :type start_time: ~datetime.datetime - :param end_time: Required. The end time for the time period for which restatement is initiated. - Only UTC time is currently supported. - :type end_time: ~datetime.datetime - :param max_concurrency: Required. The max number of parallel time windows (ready for execution) - for which a rerun is triggered. - :type max_concurrency: int - """ - - _validation = { - 'start_time': {'required': True}, - 'end_time': {'required': True}, - 'max_concurrency': {'required': True, 'maximum': 50, 'minimum': 1}, - } - - _attribute_map = { - 'start_time': {'key': 'startTime', 'type': 'iso-8601'}, - 'end_time': {'key': 'endTime', 'type': 'iso-8601'}, - 'max_concurrency': {'key': 'maxConcurrency', 'type': 'int'}, - } - - def __init__( - self, - **kwargs - ): - super(RerunTumblingWindowTriggerActionParameters, self).__init__(**kwargs) - self.start_time = kwargs['start_time'] - self.end_time = kwargs['end_time'] - self.max_concurrency = kwargs['max_concurrency'] - - -class ResponsysLinkedService(LinkedService): - """Responsys linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param endpoint: Required. The endpoint of the Responsys server. - :type endpoint: any - :param client_id: Required. The client ID associated with the Responsys application. Type: - string (or Expression with resultType string). - :type client_id: any - :param client_secret: The client secret associated with the Responsys application. Type: string - (or Expression with resultType string). - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. Type: boolean (or Expression with resultType boolean). - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. Type: boolean (or Expression with resultType boolean). - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. Type: boolean (or Expression with resultType - boolean). - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'endpoint': {'required': True}, - 'client_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ResponsysLinkedService, self).__init__(**kwargs) - self.type = 'Responsys' # type: str - self.endpoint = kwargs['endpoint'] - self.client_id = kwargs['client_id'] - self.client_secret = kwargs.get('client_secret', None) - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.use_host_verification = kwargs.get('use_host_verification', None) - self.use_peer_verification = kwargs.get('use_peer_verification', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class ResponsysObjectDataset(Dataset): - """Responsys dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ResponsysObjectDataset, self).__init__(**kwargs) - self.type = 'ResponsysObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class ResponsysSource(TabularSource): - """A copy activity Responsys source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ResponsysSource, self).__init__(**kwargs) - self.type = 'ResponsysSource' # type: str - self.query = kwargs.get('query', None) - - -class RestResourceDataset(Dataset): - """A Rest service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param relative_url: The relative URL to the resource that the RESTful API provides. Type: - string (or Expression with resultType string). - :type relative_url: any - :param request_method: The HTTP method used to call the RESTful API. The default is GET. Type: - string (or Expression with resultType string). - :type request_method: any - :param request_body: The HTTP request body to the RESTful API if requestMethod is POST. Type: - string (or Expression with resultType string). - :type request_body: any - :param additional_headers: The additional HTTP headers in the request to the RESTful API. Type: - string (or Expression with resultType string). - :type additional_headers: any - :param pagination_rules: The pagination rules to compose next page requests. Type: string (or - Expression with resultType string). - :type pagination_rules: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'relative_url': {'key': 'typeProperties.relativeUrl', 'type': 'object'}, - 'request_method': {'key': 'typeProperties.requestMethod', 'type': 'object'}, - 'request_body': {'key': 'typeProperties.requestBody', 'type': 'object'}, - 'additional_headers': {'key': 'typeProperties.additionalHeaders', 'type': 'object'}, - 'pagination_rules': {'key': 'typeProperties.paginationRules', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(RestResourceDataset, self).__init__(**kwargs) - self.type = 'RestResource' # type: str - self.relative_url = kwargs.get('relative_url', None) - self.request_method = kwargs.get('request_method', None) - self.request_body = kwargs.get('request_body', None) - self.additional_headers = kwargs.get('additional_headers', None) - self.pagination_rules = kwargs.get('pagination_rules', None) - - -class RestServiceLinkedService(LinkedService): - """Rest Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param url: Required. The base URL of the REST service. - :type url: any - :param enable_server_certificate_validation: Whether to validate server side SSL certificate - when connecting to the endpoint.The default value is true. Type: boolean (or Expression with - resultType boolean). - :type enable_server_certificate_validation: any - :param authentication_type: Required. Type of authentication used to connect to the REST - service. Possible values include: "Anonymous", "Basic", "AadServicePrincipal", - "ManagedServiceIdentity". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.RestServiceAuthenticationType - :param user_name: The user name used in Basic authentication type. - :type user_name: any - :param password: The password used in Basic authentication type. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param service_principal_id: The application's client ID used in AadServicePrincipal - authentication type. - :type service_principal_id: any - :param service_principal_key: The application's key used in AadServicePrincipal authentication - type. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: The tenant information (domain name or tenant ID) used in AadServicePrincipal - authentication type under which your application resides. - :type tenant: any - :param azure_cloud_type: Indicates the azure cloud type of the service principle auth. Allowed - values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data - factory regions’ cloud type. Type: string (or Expression with resultType string). - :type azure_cloud_type: any - :param aad_resource_id: The resource you are requesting authorization to use. - :type aad_resource_id: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'enable_server_certificate_validation': {'key': 'typeProperties.enableServerCertificateValidation', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'azure_cloud_type': {'key': 'typeProperties.azureCloudType', 'type': 'object'}, - 'aad_resource_id': {'key': 'typeProperties.aadResourceId', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(RestServiceLinkedService, self).__init__(**kwargs) - self.type = 'RestService' # type: str - self.url = kwargs['url'] - self.enable_server_certificate_validation = kwargs.get('enable_server_certificate_validation', None) - self.authentication_type = kwargs['authentication_type'] - self.user_name = kwargs.get('user_name', None) - self.password = kwargs.get('password', None) - self.service_principal_id = kwargs.get('service_principal_id', None) - self.service_principal_key = kwargs.get('service_principal_key', None) - self.tenant = kwargs.get('tenant', None) - self.azure_cloud_type = kwargs.get('azure_cloud_type', None) - self.aad_resource_id = kwargs.get('aad_resource_id', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class RestSink(CopySink): - """A copy activity Rest service Sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param request_method: The HTTP method used to call the RESTful API. The default is POST. Type: - string (or Expression with resultType string). - :type request_method: any - :param additional_headers: The additional HTTP headers in the request to the RESTful API. Type: - string (or Expression with resultType string). - :type additional_headers: any - :param http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout - to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: any - :param request_interval: The time to await before sending next request, in milliseconds. - :type request_interval: any - :param http_compression_type: Http Compression Type to Send data in compressed format with - Optimal Compression Level, Default is None. And The Only Supported option is Gzip. - :type http_compression_type: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'request_method': {'key': 'requestMethod', 'type': 'object'}, - 'additional_headers': {'key': 'additionalHeaders', 'type': 'object'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - 'request_interval': {'key': 'requestInterval', 'type': 'object'}, - 'http_compression_type': {'key': 'httpCompressionType', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(RestSink, self).__init__(**kwargs) - self.type = 'RestSink' # type: str - self.request_method = kwargs.get('request_method', None) - self.additional_headers = kwargs.get('additional_headers', None) - self.http_request_timeout = kwargs.get('http_request_timeout', None) - self.request_interval = kwargs.get('request_interval', None) - self.http_compression_type = kwargs.get('http_compression_type', None) - - -class RestSource(CopySource): - """A copy activity Rest service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param request_method: The HTTP method used to call the RESTful API. The default is GET. Type: - string (or Expression with resultType string). - :type request_method: any - :param request_body: The HTTP request body to the RESTful API if requestMethod is POST. Type: - string (or Expression with resultType string). - :type request_body: any - :param additional_headers: The additional HTTP headers in the request to the RESTful API. Type: - string (or Expression with resultType string). - :type additional_headers: any - :param pagination_rules: The pagination rules to compose next page requests. Type: string (or - Expression with resultType string). - :type pagination_rules: any - :param http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout - to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: any - :param request_interval: The time to await before sending next page request. - :type request_interval: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'request_method': {'key': 'requestMethod', 'type': 'object'}, - 'request_body': {'key': 'requestBody', 'type': 'object'}, - 'additional_headers': {'key': 'additionalHeaders', 'type': 'object'}, - 'pagination_rules': {'key': 'paginationRules', 'type': 'object'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - 'request_interval': {'key': 'requestInterval', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(RestSource, self).__init__(**kwargs) - self.type = 'RestSource' # type: str - self.request_method = kwargs.get('request_method', None) - self.request_body = kwargs.get('request_body', None) - self.additional_headers = kwargs.get('additional_headers', None) - self.pagination_rules = kwargs.get('pagination_rules', None) - self.http_request_timeout = kwargs.get('http_request_timeout', None) - self.request_interval = kwargs.get('request_interval', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class RetryPolicy(msrest.serialization.Model): - """Execution policy for an activity. - - :param count: Maximum ordinary retry attempts. Default is 0. Type: integer (or Expression with - resultType integer), minimum: 0. - :type count: any - :param interval_in_seconds: Interval between retries in seconds. Default is 30. - :type interval_in_seconds: int - """ - - _validation = { - 'interval_in_seconds': {'maximum': 86400, 'minimum': 30}, - } - - _attribute_map = { - 'count': {'key': 'count', 'type': 'object'}, - 'interval_in_seconds': {'key': 'intervalInSeconds', 'type': 'int'}, - } - - def __init__( - self, - **kwargs - ): - super(RetryPolicy, self).__init__(**kwargs) - self.count = kwargs.get('count', None) - self.interval_in_seconds = kwargs.get('interval_in_seconds', None) - - -class RunFilterParameters(msrest.serialization.Model): - """Query parameters for listing runs. - - All required parameters must be populated in order to send to Azure. - - :param continuation_token: The continuation token for getting the next page of results. Null - for first page. - :type continuation_token: str - :param last_updated_after: Required. The time at or after which the run event was updated in - 'ISO 8601' format. - :type last_updated_after: ~datetime.datetime - :param last_updated_before: Required. The time at or before which the run event was updated in - 'ISO 8601' format. - :type last_updated_before: ~datetime.datetime - :param filters: List of filters. - :type filters: list[~azure.synapse.artifacts.v2020_12_01.models.RunQueryFilter] - :param order_by: List of OrderBy option. - :type order_by: list[~azure.synapse.artifacts.v2020_12_01.models.RunQueryOrderBy] - """ - - _validation = { - 'last_updated_after': {'required': True}, - 'last_updated_before': {'required': True}, - } - - _attribute_map = { - 'continuation_token': {'key': 'continuationToken', 'type': 'str'}, - 'last_updated_after': {'key': 'lastUpdatedAfter', 'type': 'iso-8601'}, - 'last_updated_before': {'key': 'lastUpdatedBefore', 'type': 'iso-8601'}, - 'filters': {'key': 'filters', 'type': '[RunQueryFilter]'}, - 'order_by': {'key': 'orderBy', 'type': '[RunQueryOrderBy]'}, - } - - def __init__( - self, - **kwargs - ): - super(RunFilterParameters, self).__init__(**kwargs) - self.continuation_token = kwargs.get('continuation_token', None) - self.last_updated_after = kwargs['last_updated_after'] - self.last_updated_before = kwargs['last_updated_before'] - self.filters = kwargs.get('filters', None) - self.order_by = kwargs.get('order_by', None) - - -class RunQueryFilter(msrest.serialization.Model): - """Query filter option for listing runs. - - All required parameters must be populated in order to send to Azure. - - :param operand: Required. Parameter name to be used for filter. The allowed operands to query - pipeline runs are PipelineName, RunStart, RunEnd and Status; to query activity runs are - ActivityName, ActivityRunStart, ActivityRunEnd, ActivityType and Status, and to query trigger - runs are TriggerName, TriggerRunTimestamp and Status. Possible values include: "PipelineName", - "Status", "RunStart", "RunEnd", "ActivityName", "ActivityRunStart", "ActivityRunEnd", - "ActivityType", "TriggerName", "TriggerRunTimestamp", "RunGroupId", "LatestOnly". - :type operand: str or ~azure.synapse.artifacts.v2020_12_01.models.RunQueryFilterOperand - :param operator: Required. Operator to be used for filter. Possible values include: "Equals", - "NotEquals", "In", "NotIn". - :type operator: str or ~azure.synapse.artifacts.v2020_12_01.models.RunQueryFilterOperator - :param values: Required. List of filter values. - :type values: list[str] - """ - - _validation = { - 'operand': {'required': True}, - 'operator': {'required': True}, - 'values': {'required': True}, - } - - _attribute_map = { - 'operand': {'key': 'operand', 'type': 'str'}, - 'operator': {'key': 'operator', 'type': 'str'}, - 'values': {'key': 'values', 'type': '[str]'}, - } - - def __init__( - self, - **kwargs - ): - super(RunQueryFilter, self).__init__(**kwargs) - self.operand = kwargs['operand'] - self.operator = kwargs['operator'] - self.values = kwargs['values'] - - -class RunQueryOrderBy(msrest.serialization.Model): - """An object to provide order by options for listing runs. - - All required parameters must be populated in order to send to Azure. - - :param order_by: Required. Parameter name to be used for order by. The allowed parameters to - order by for pipeline runs are PipelineName, RunStart, RunEnd and Status; for activity runs are - ActivityName, ActivityRunStart, ActivityRunEnd and Status; for trigger runs are TriggerName, - TriggerRunTimestamp and Status. Possible values include: "RunStart", "RunEnd", "PipelineName", - "Status", "ActivityName", "ActivityRunStart", "ActivityRunEnd", "TriggerName", - "TriggerRunTimestamp". - :type order_by: str or ~azure.synapse.artifacts.v2020_12_01.models.RunQueryOrderByField - :param order: Required. Sorting order of the parameter. Possible values include: "ASC", "DESC". - :type order: str or ~azure.synapse.artifacts.v2020_12_01.models.RunQueryOrder - """ - - _validation = { - 'order_by': {'required': True}, - 'order': {'required': True}, - } - - _attribute_map = { - 'order_by': {'key': 'orderBy', 'type': 'str'}, - 'order': {'key': 'order', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(RunQueryOrderBy, self).__init__(**kwargs) - self.order_by = kwargs['order_by'] - self.order = kwargs['order'] - - -class SalesforceLinkedService(LinkedService): - """Linked service for Salesforce. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param environment_url: The URL of Salesforce instance. Default is - 'https://login.salesforce.com'. To copy data from sandbox, specify - 'https://test.salesforce.com'. To copy data from custom domain, specify, for example, - 'https://[domain].my.salesforce.com'. Type: string (or Expression with resultType string). - :type environment_url: any - :param username: The username for Basic authentication of the Salesforce instance. Type: string - (or Expression with resultType string). - :type username: any - :param password: The password for Basic authentication of the Salesforce instance. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param security_token: The security token is optional to remotely access Salesforce instance. - :type security_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param api_version: The Salesforce API version used in ADF. Type: string (or Expression with - resultType string). - :type api_version: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'environment_url': {'key': 'typeProperties.environmentUrl', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'security_token': {'key': 'typeProperties.securityToken', 'type': 'SecretBase'}, - 'api_version': {'key': 'typeProperties.apiVersion', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SalesforceLinkedService, self).__init__(**kwargs) - self.type = 'Salesforce' # type: str - self.environment_url = kwargs.get('environment_url', None) - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.security_token = kwargs.get('security_token', None) - self.api_version = kwargs.get('api_version', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class SalesforceMarketingCloudLinkedService(LinkedService): - """Salesforce Marketing Cloud linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_properties: Properties used to connect to Salesforce Marketing Cloud. It is - mutually exclusive with any other properties in the linked service. Type: object. - :type connection_properties: any - :param client_id: Required. The client ID associated with the Salesforce Marketing Cloud - application. Type: string (or Expression with resultType string). - :type client_id: any - :param client_secret: The client secret associated with the Salesforce Marketing Cloud - application. Type: string (or Expression with resultType string). - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. Type: boolean (or Expression with resultType boolean). - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. Type: boolean (or Expression with resultType boolean). - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. Type: boolean (or Expression with resultType - boolean). - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'client_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_properties': {'key': 'typeProperties.connectionProperties', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SalesforceMarketingCloudLinkedService, self).__init__(**kwargs) - self.type = 'SalesforceMarketingCloud' # type: str - self.connection_properties = kwargs.get('connection_properties', None) - self.client_id = kwargs['client_id'] - self.client_secret = kwargs.get('client_secret', None) - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.use_host_verification = kwargs.get('use_host_verification', None) - self.use_peer_verification = kwargs.get('use_peer_verification', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class SalesforceMarketingCloudObjectDataset(Dataset): - """Salesforce Marketing Cloud dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SalesforceMarketingCloudObjectDataset, self).__init__(**kwargs) - self.type = 'SalesforceMarketingCloudObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class SalesforceMarketingCloudSource(TabularSource): - """A copy activity Salesforce Marketing Cloud source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SalesforceMarketingCloudSource, self).__init__(**kwargs) - self.type = 'SalesforceMarketingCloudSource' # type: str - self.query = kwargs.get('query', None) - - -class SalesforceObjectDataset(Dataset): - """The Salesforce object dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param object_api_name: The Salesforce object API name. Type: string (or Expression with - resultType string). - :type object_api_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'object_api_name': {'key': 'typeProperties.objectApiName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SalesforceObjectDataset, self).__init__(**kwargs) - self.type = 'SalesforceObject' # type: str - self.object_api_name = kwargs.get('object_api_name', None) - - -class SalesforceServiceCloudLinkedService(LinkedService): - """Linked service for Salesforce Service Cloud. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param environment_url: The URL of Salesforce Service Cloud instance. Default is - 'https://login.salesforce.com'. To copy data from sandbox, specify - 'https://test.salesforce.com'. To copy data from custom domain, specify, for example, - 'https://[domain].my.salesforce.com'. Type: string (or Expression with resultType string). - :type environment_url: any - :param username: The username for Basic authentication of the Salesforce instance. Type: string - (or Expression with resultType string). - :type username: any - :param password: The password for Basic authentication of the Salesforce instance. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param security_token: The security token is optional to remotely access Salesforce instance. - :type security_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param api_version: The Salesforce API version used in ADF. Type: string (or Expression with - resultType string). - :type api_version: any - :param extended_properties: Extended properties appended to the connection string. Type: string - (or Expression with resultType string). - :type extended_properties: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'environment_url': {'key': 'typeProperties.environmentUrl', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'security_token': {'key': 'typeProperties.securityToken', 'type': 'SecretBase'}, - 'api_version': {'key': 'typeProperties.apiVersion', 'type': 'object'}, - 'extended_properties': {'key': 'typeProperties.extendedProperties', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SalesforceServiceCloudLinkedService, self).__init__(**kwargs) - self.type = 'SalesforceServiceCloud' # type: str - self.environment_url = kwargs.get('environment_url', None) - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.security_token = kwargs.get('security_token', None) - self.api_version = kwargs.get('api_version', None) - self.extended_properties = kwargs.get('extended_properties', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class SalesforceServiceCloudObjectDataset(Dataset): - """The Salesforce Service Cloud object dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param object_api_name: The Salesforce Service Cloud object API name. Type: string (or - Expression with resultType string). - :type object_api_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'object_api_name': {'key': 'typeProperties.objectApiName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SalesforceServiceCloudObjectDataset, self).__init__(**kwargs) - self.type = 'SalesforceServiceCloudObject' # type: str - self.object_api_name = kwargs.get('object_api_name', None) - - -class SalesforceServiceCloudSink(CopySink): - """A copy activity Salesforce Service Cloud sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param write_behavior: The write behavior for the operation. Default is Insert. Possible values - include: "Insert", "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2020_12_01.models.SalesforceSinkWriteBehavior - :param external_id_field_name: The name of the external ID field for upsert operation. Default - value is 'Id' column. Type: string (or Expression with resultType string). - :type external_id_field_name: any - :param ignore_null_values: The flag indicating whether or not to ignore null values from input - dataset (except key fields) during write operation. Default value is false. If set it to true, - it means ADF will leave the data in the destination object unchanged when doing upsert/update - operation and insert defined default value when doing insert operation, versus ADF will update - the data in the destination object to NULL when doing upsert/update operation and insert NULL - value when doing insert operation. Type: boolean (or Expression with resultType boolean). - :type ignore_null_values: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, - 'external_id_field_name': {'key': 'externalIdFieldName', 'type': 'object'}, - 'ignore_null_values': {'key': 'ignoreNullValues', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SalesforceServiceCloudSink, self).__init__(**kwargs) - self.type = 'SalesforceServiceCloudSink' # type: str - self.write_behavior = kwargs.get('write_behavior', None) - self.external_id_field_name = kwargs.get('external_id_field_name', None) - self.ignore_null_values = kwargs.get('ignore_null_values', None) - - -class SalesforceServiceCloudSource(CopySource): - """A copy activity Salesforce Service Cloud source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - :param read_behavior: The read behavior for the operation. Default is Query. Possible values - include: "Query", "QueryAll". - :type read_behavior: str or - ~azure.synapse.artifacts.v2020_12_01.models.SalesforceSourceReadBehavior - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'read_behavior': {'key': 'readBehavior', 'type': 'str'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SalesforceServiceCloudSource, self).__init__(**kwargs) - self.type = 'SalesforceServiceCloudSource' # type: str - self.query = kwargs.get('query', None) - self.read_behavior = kwargs.get('read_behavior', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class SalesforceSink(CopySink): - """A copy activity Salesforce sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param write_behavior: The write behavior for the operation. Default is Insert. Possible values - include: "Insert", "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2020_12_01.models.SalesforceSinkWriteBehavior - :param external_id_field_name: The name of the external ID field for upsert operation. Default - value is 'Id' column. Type: string (or Expression with resultType string). - :type external_id_field_name: any - :param ignore_null_values: The flag indicating whether or not to ignore null values from input - dataset (except key fields) during write operation. Default value is false. If set it to true, - it means ADF will leave the data in the destination object unchanged when doing upsert/update - operation and insert defined default value when doing insert operation, versus ADF will update - the data in the destination object to NULL when doing upsert/update operation and insert NULL - value when doing insert operation. Type: boolean (or Expression with resultType boolean). - :type ignore_null_values: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, - 'external_id_field_name': {'key': 'externalIdFieldName', 'type': 'object'}, - 'ignore_null_values': {'key': 'ignoreNullValues', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SalesforceSink, self).__init__(**kwargs) - self.type = 'SalesforceSink' # type: str - self.write_behavior = kwargs.get('write_behavior', None) - self.external_id_field_name = kwargs.get('external_id_field_name', None) - self.ignore_null_values = kwargs.get('ignore_null_values', None) - - -class SalesforceSource(TabularSource): - """A copy activity Salesforce source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - :param read_behavior: The read behavior for the operation. Default is Query. Possible values - include: "Query", "QueryAll". - :type read_behavior: str or - ~azure.synapse.artifacts.v2020_12_01.models.SalesforceSourceReadBehavior - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'read_behavior': {'key': 'readBehavior', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SalesforceSource, self).__init__(**kwargs) - self.type = 'SalesforceSource' # type: str - self.query = kwargs.get('query', None) - self.read_behavior = kwargs.get('read_behavior', None) - - -class SapBwCubeDataset(Dataset): - """The SAP BW cube dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - } - - def __init__( - self, - **kwargs - ): - super(SapBwCubeDataset, self).__init__(**kwargs) - self.type = 'SapBwCube' # type: str - - -class SapBWLinkedService(LinkedService): - """SAP Business Warehouse Linked Service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param server: Required. Host name of the SAP BW instance. Type: string (or Expression with - resultType string). - :type server: any - :param system_number: Required. System number of the BW system. (Usually a two-digit decimal - number represented as a string.) Type: string (or Expression with resultType string). - :type system_number: any - :param client_id: Required. Client ID of the client on the BW system. (Usually a three-digit - decimal number represented as a string) Type: string (or Expression with resultType string). - :type client_id: any - :param user_name: Username to access the SAP BW server. Type: string (or Expression with - resultType string). - :type user_name: any - :param password: Password to access the SAP BW server. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'server': {'required': True}, - 'system_number': {'required': True}, - 'client_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'system_number': {'key': 'typeProperties.systemNumber', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SapBWLinkedService, self).__init__(**kwargs) - self.type = 'SapBW' # type: str - self.server = kwargs['server'] - self.system_number = kwargs['system_number'] - self.client_id = kwargs['client_id'] - self.user_name = kwargs.get('user_name', None) - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class SapBwSource(TabularSource): - """A copy activity source for SapBW server via MDX. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: MDX query. Type: string (or Expression with resultType string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SapBwSource, self).__init__(**kwargs) - self.type = 'SapBwSource' # type: str - self.query = kwargs.get('query', None) - - -class SapCloudForCustomerLinkedService(LinkedService): - """Linked service for SAP Cloud for Customer. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param url: Required. The URL of SAP Cloud for Customer OData API. For example, - '[https://[tenantname].crm.ondemand.com/sap/c4c/odata/v1]'. Type: string (or Expression with - resultType string). - :type url: any - :param username: The username for Basic authentication. Type: string (or Expression with - resultType string). - :type username: any - :param password: The password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Either encryptedCredential or - username/password must be provided. Type: string (or Expression with resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SapCloudForCustomerLinkedService, self).__init__(**kwargs) - self.type = 'SapCloudForCustomer' # type: str - self.url = kwargs['url'] - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class SapCloudForCustomerResourceDataset(Dataset): - """The path of the SAP Cloud for Customer OData entity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param path: Required. The path of the SAP Cloud for Customer OData entity. Type: string (or - Expression with resultType string). - :type path: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'path': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'path': {'key': 'typeProperties.path', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SapCloudForCustomerResourceDataset, self).__init__(**kwargs) - self.type = 'SapCloudForCustomerResource' # type: str - self.path = kwargs['path'] - - -class SapCloudForCustomerSink(CopySink): - """A copy activity SAP Cloud for Customer sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param write_behavior: The write behavior for the operation. Default is 'Insert'. Possible - values include: "Insert", "Update". - :type write_behavior: str or - ~azure.synapse.artifacts.v2020_12_01.models.SapCloudForCustomerSinkWriteBehavior - :param http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout - to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SapCloudForCustomerSink, self).__init__(**kwargs) - self.type = 'SapCloudForCustomerSink' # type: str - self.write_behavior = kwargs.get('write_behavior', None) - self.http_request_timeout = kwargs.get('http_request_timeout', None) - - -class SapCloudForCustomerSource(TabularSource): - """A copy activity source for SAP Cloud for Customer source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: SAP Cloud for Customer OData query. For example, "$top=1". Type: string (or - Expression with resultType string). - :type query: any - :param http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout - to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SapCloudForCustomerSource, self).__init__(**kwargs) - self.type = 'SapCloudForCustomerSource' # type: str - self.query = kwargs.get('query', None) - self.http_request_timeout = kwargs.get('http_request_timeout', None) - - -class SapEccLinkedService(LinkedService): - """Linked service for SAP ERP Central Component(SAP ECC). - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param url: Required. The URL of SAP ECC OData API. For example, - '[https://hostname:port/sap/opu/odata/sap/servicename/]'. Type: string (or Expression with - resultType string). - :type url: str - :param username: The username for Basic authentication. Type: string (or Expression with - resultType string). - :type username: str - :param password: The password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Either encryptedCredential or - username/password must be provided. Type: string (or Expression with resultType string). - :type encrypted_credential: str - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'url': {'key': 'typeProperties.url', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'str'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SapEccLinkedService, self).__init__(**kwargs) - self.type = 'SapEcc' # type: str - self.url = kwargs['url'] - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class SapEccResourceDataset(Dataset): - """The path of the SAP ECC OData entity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param path: Required. The path of the SAP ECC OData entity. Type: string (or Expression with - resultType string). - :type path: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'path': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'path': {'key': 'typeProperties.path', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SapEccResourceDataset, self).__init__(**kwargs) - self.type = 'SapEccResource' # type: str - self.path = kwargs['path'] - - -class SapEccSource(TabularSource): - """A copy activity source for SAP ECC source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: SAP ECC OData query. For example, "$top=1". Type: string (or Expression with - resultType string). - :type query: any - :param http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout - to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SapEccSource, self).__init__(**kwargs) - self.type = 'SapEccSource' # type: str - self.query = kwargs.get('query', None) - self.http_request_timeout = kwargs.get('http_request_timeout', None) - - -class SapHanaLinkedService(LinkedService): - """SAP HANA Linked Service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: SAP HANA ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param server: Required. Host name of the SAP HANA server. Type: string (or Expression with - resultType string). - :type server: any - :param authentication_type: The authentication type to be used to connect to the SAP HANA - server. Possible values include: "Basic", "Windows". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.SapHanaAuthenticationType - :param user_name: Username to access the SAP HANA server. Type: string (or Expression with - resultType string). - :type user_name: any - :param password: Password to access the SAP HANA server. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'server': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SapHanaLinkedService, self).__init__(**kwargs) - self.type = 'SapHana' # type: str - self.connection_string = kwargs.get('connection_string', None) - self.server = kwargs['server'] - self.authentication_type = kwargs.get('authentication_type', None) - self.user_name = kwargs.get('user_name', None) - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class SapHanaPartitionSettings(msrest.serialization.Model): - """The settings that will be leveraged for SAP HANA source partitioning. - - :param partition_column_name: The name of the column that will be used for proceeding range - partitioning. Type: string (or Expression with resultType string). - :type partition_column_name: any - """ - - _attribute_map = { - 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SapHanaPartitionSettings, self).__init__(**kwargs) - self.partition_column_name = kwargs.get('partition_column_name', None) - - -class SapHanaSource(TabularSource): - """A copy activity source for SAP HANA source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: SAP HANA Sql query. Type: string (or Expression with resultType string). - :type query: any - :param packet_size: The packet size of data read from SAP HANA. Type: integer(or Expression - with resultType integer). - :type packet_size: any - :param partition_option: The partition mechanism that will be used for SAP HANA read in - parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "SapHanaDynamicRange". - :type partition_option: str or - ~azure.synapse.artifacts.v2020_12_01.models.SapHanaPartitionOption - :param partition_settings: The settings that will be leveraged for SAP HANA source - partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.SapHanaPartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'packet_size': {'key': 'packetSize', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'SapHanaPartitionSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(SapHanaSource, self).__init__(**kwargs) - self.type = 'SapHanaSource' # type: str - self.query = kwargs.get('query', None) - self.packet_size = kwargs.get('packet_size', None) - self.partition_option = kwargs.get('partition_option', None) - self.partition_settings = kwargs.get('partition_settings', None) - - -class SapHanaTableDataset(Dataset): - """SAP HANA Table properties. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param schema_type_properties_schema: The schema name of SAP HANA. Type: string (or Expression - with resultType string). - :type schema_type_properties_schema: any - :param table: The table name of SAP HANA. Type: string (or Expression with resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SapHanaTableDataset, self).__init__(**kwargs) - self.type = 'SapHanaTable' # type: str - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - self.table = kwargs.get('table', None) - - -class SapOpenHubLinkedService(LinkedService): - """SAP Business Warehouse Open Hub Destination Linked Service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param server: Required. Host name of the SAP BW instance where the open hub destination is - located. Type: string (or Expression with resultType string). - :type server: any - :param system_number: Required. System number of the BW system where the open hub destination - is located. (Usually a two-digit decimal number represented as a string.) Type: string (or - Expression with resultType string). - :type system_number: any - :param client_id: Required. Client ID of the client on the BW system where the open hub - destination is located. (Usually a three-digit decimal number represented as a string) Type: - string (or Expression with resultType string). - :type client_id: any - :param language: Language of the BW system where the open hub destination is located. The - default value is EN. Type: string (or Expression with resultType string). - :type language: any - :param system_id: SystemID of the SAP system where the table is located. Type: string (or - Expression with resultType string). - :type system_id: any - :param user_name: Username to access the SAP BW server where the open hub destination is - located. Type: string (or Expression with resultType string). - :type user_name: any - :param password: Password to access the SAP BW server where the open hub destination is - located. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param message_server: The hostname of the SAP Message Server. Type: string (or Expression with - resultType string). - :type message_server: any - :param message_server_service: The service name or port number of the Message Server. Type: - string (or Expression with resultType string). - :type message_server_service: any - :param logon_group: The Logon Group for the SAP System. Type: string (or Expression with - resultType string). - :type logon_group: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'server': {'required': True}, - 'system_number': {'required': True}, - 'client_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'system_number': {'key': 'typeProperties.systemNumber', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'language': {'key': 'typeProperties.language', 'type': 'object'}, - 'system_id': {'key': 'typeProperties.systemId', 'type': 'object'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'message_server': {'key': 'typeProperties.messageServer', 'type': 'object'}, - 'message_server_service': {'key': 'typeProperties.messageServerService', 'type': 'object'}, - 'logon_group': {'key': 'typeProperties.logonGroup', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SapOpenHubLinkedService, self).__init__(**kwargs) - self.type = 'SapOpenHub' # type: str - self.server = kwargs['server'] - self.system_number = kwargs['system_number'] - self.client_id = kwargs['client_id'] - self.language = kwargs.get('language', None) - self.system_id = kwargs.get('system_id', None) - self.user_name = kwargs.get('user_name', None) - self.password = kwargs.get('password', None) - self.message_server = kwargs.get('message_server', None) - self.message_server_service = kwargs.get('message_server_service', None) - self.logon_group = kwargs.get('logon_group', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class SapOpenHubSource(TabularSource): - """A copy activity source for SAP Business Warehouse Open Hub Destination source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param exclude_last_request: Whether to exclude the records of the last request. The default - value is true. Type: boolean (or Expression with resultType boolean). - :type exclude_last_request: any - :param base_request_id: The ID of request for delta loading. Once it is set, only data with - requestId larger than the value of this property will be retrieved. The default value is 0. - Type: integer (or Expression with resultType integer ). - :type base_request_id: any - :param custom_rfc_read_table_function_module: Specifies the custom RFC function module that - will be used to read data from SAP Table. Type: string (or Expression with resultType string). - :type custom_rfc_read_table_function_module: any - :param sap_data_column_delimiter: The single character that will be used as delimiter passed to - SAP RFC as well as splitting the output data retrieved. Type: string (or Expression with - resultType string). - :type sap_data_column_delimiter: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'exclude_last_request': {'key': 'excludeLastRequest', 'type': 'object'}, - 'base_request_id': {'key': 'baseRequestId', 'type': 'object'}, - 'custom_rfc_read_table_function_module': {'key': 'customRfcReadTableFunctionModule', 'type': 'object'}, - 'sap_data_column_delimiter': {'key': 'sapDataColumnDelimiter', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SapOpenHubSource, self).__init__(**kwargs) - self.type = 'SapOpenHubSource' # type: str - self.exclude_last_request = kwargs.get('exclude_last_request', None) - self.base_request_id = kwargs.get('base_request_id', None) - self.custom_rfc_read_table_function_module = kwargs.get('custom_rfc_read_table_function_module', None) - self.sap_data_column_delimiter = kwargs.get('sap_data_column_delimiter', None) - - -class SapOpenHubTableDataset(Dataset): - """Sap Business Warehouse Open Hub Destination Table properties. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param open_hub_destination_name: Required. The name of the Open Hub Destination with - destination type as Database Table. Type: string (or Expression with resultType string). - :type open_hub_destination_name: any - :param exclude_last_request: Whether to exclude the records of the last request. The default - value is true. Type: boolean (or Expression with resultType boolean). - :type exclude_last_request: any - :param base_request_id: The ID of request for delta loading. Once it is set, only data with - requestId larger than the value of this property will be retrieved. The default value is 0. - Type: integer (or Expression with resultType integer ). - :type base_request_id: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'open_hub_destination_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'open_hub_destination_name': {'key': 'typeProperties.openHubDestinationName', 'type': 'object'}, - 'exclude_last_request': {'key': 'typeProperties.excludeLastRequest', 'type': 'object'}, - 'base_request_id': {'key': 'typeProperties.baseRequestId', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SapOpenHubTableDataset, self).__init__(**kwargs) - self.type = 'SapOpenHubTable' # type: str - self.open_hub_destination_name = kwargs['open_hub_destination_name'] - self.exclude_last_request = kwargs.get('exclude_last_request', None) - self.base_request_id = kwargs.get('base_request_id', None) - - -class SapTableLinkedService(LinkedService): - """SAP Table Linked Service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param server: Host name of the SAP instance where the table is located. Type: string (or - Expression with resultType string). - :type server: any - :param system_number: System number of the SAP system where the table is located. (Usually a - two-digit decimal number represented as a string.) Type: string (or Expression with resultType - string). - :type system_number: any - :param client_id: Client ID of the client on the SAP system where the table is located. - (Usually a three-digit decimal number represented as a string) Type: string (or Expression with - resultType string). - :type client_id: any - :param language: Language of the SAP system where the table is located. The default value is - EN. Type: string (or Expression with resultType string). - :type language: any - :param system_id: SystemID of the SAP system where the table is located. Type: string (or - Expression with resultType string). - :type system_id: any - :param user_name: Username to access the SAP server where the table is located. Type: string - (or Expression with resultType string). - :type user_name: any - :param password: Password to access the SAP server where the table is located. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param message_server: The hostname of the SAP Message Server. Type: string (or Expression with - resultType string). - :type message_server: any - :param message_server_service: The service name or port number of the Message Server. Type: - string (or Expression with resultType string). - :type message_server_service: any - :param snc_mode: SNC activation indicator to access the SAP server where the table is located. - Must be either 0 (off) or 1 (on). Type: string (or Expression with resultType string). - :type snc_mode: any - :param snc_my_name: Initiator's SNC name to access the SAP server where the table is located. - Type: string (or Expression with resultType string). - :type snc_my_name: any - :param snc_partner_name: Communication partner's SNC name to access the SAP server where the - table is located. Type: string (or Expression with resultType string). - :type snc_partner_name: any - :param snc_library_path: External security product's library to access the SAP server where the - table is located. Type: string (or Expression with resultType string). - :type snc_library_path: any - :param snc_qop: SNC Quality of Protection. Allowed value include: 1, 2, 3, 8, 9. Type: string - (or Expression with resultType string). - :type snc_qop: any - :param logon_group: The Logon Group for the SAP System. Type: string (or Expression with - resultType string). - :type logon_group: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'system_number': {'key': 'typeProperties.systemNumber', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'language': {'key': 'typeProperties.language', 'type': 'object'}, - 'system_id': {'key': 'typeProperties.systemId', 'type': 'object'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'message_server': {'key': 'typeProperties.messageServer', 'type': 'object'}, - 'message_server_service': {'key': 'typeProperties.messageServerService', 'type': 'object'}, - 'snc_mode': {'key': 'typeProperties.sncMode', 'type': 'object'}, - 'snc_my_name': {'key': 'typeProperties.sncMyName', 'type': 'object'}, - 'snc_partner_name': {'key': 'typeProperties.sncPartnerName', 'type': 'object'}, - 'snc_library_path': {'key': 'typeProperties.sncLibraryPath', 'type': 'object'}, - 'snc_qop': {'key': 'typeProperties.sncQop', 'type': 'object'}, - 'logon_group': {'key': 'typeProperties.logonGroup', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SapTableLinkedService, self).__init__(**kwargs) - self.type = 'SapTable' # type: str - self.server = kwargs.get('server', None) - self.system_number = kwargs.get('system_number', None) - self.client_id = kwargs.get('client_id', None) - self.language = kwargs.get('language', None) - self.system_id = kwargs.get('system_id', None) - self.user_name = kwargs.get('user_name', None) - self.password = kwargs.get('password', None) - self.message_server = kwargs.get('message_server', None) - self.message_server_service = kwargs.get('message_server_service', None) - self.snc_mode = kwargs.get('snc_mode', None) - self.snc_my_name = kwargs.get('snc_my_name', None) - self.snc_partner_name = kwargs.get('snc_partner_name', None) - self.snc_library_path = kwargs.get('snc_library_path', None) - self.snc_qop = kwargs.get('snc_qop', None) - self.logon_group = kwargs.get('logon_group', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class SapTablePartitionSettings(msrest.serialization.Model): - """The settings that will be leveraged for SAP table source partitioning. - - :param partition_column_name: The name of the column that will be used for proceeding range - partitioning. Type: string (or Expression with resultType string). - :type partition_column_name: any - :param partition_upper_bound: The maximum value of column specified in partitionColumnName that - will be used for proceeding range partitioning. Type: string (or Expression with resultType - string). - :type partition_upper_bound: any - :param partition_lower_bound: The minimum value of column specified in partitionColumnName that - will be used for proceeding range partitioning. Type: string (or Expression with resultType - string). - :type partition_lower_bound: any - :param max_partitions_number: The maximum value of partitions the table will be split into. - Type: integer (or Expression with resultType string). - :type max_partitions_number: any - """ - - _attribute_map = { - 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, - 'partition_upper_bound': {'key': 'partitionUpperBound', 'type': 'object'}, - 'partition_lower_bound': {'key': 'partitionLowerBound', 'type': 'object'}, - 'max_partitions_number': {'key': 'maxPartitionsNumber', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SapTablePartitionSettings, self).__init__(**kwargs) - self.partition_column_name = kwargs.get('partition_column_name', None) - self.partition_upper_bound = kwargs.get('partition_upper_bound', None) - self.partition_lower_bound = kwargs.get('partition_lower_bound', None) - self.max_partitions_number = kwargs.get('max_partitions_number', None) - - -class SapTableResourceDataset(Dataset): - """SAP Table Resource properties. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: Required. The name of the SAP Table. Type: string (or Expression with - resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'table_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SapTableResourceDataset, self).__init__(**kwargs) - self.type = 'SapTableResource' # type: str - self.table_name = kwargs['table_name'] - - -class SapTableSource(TabularSource): - """A copy activity source for SAP Table source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param row_count: The number of rows to be retrieved. Type: integer(or Expression with - resultType integer). - :type row_count: any - :param row_skips: The number of rows that will be skipped. Type: integer (or Expression with - resultType integer). - :type row_skips: any - :param rfc_table_fields: The fields of the SAP table that will be retrieved. For example, - column0, column1. Type: string (or Expression with resultType string). - :type rfc_table_fields: any - :param rfc_table_options: The options for the filtering of the SAP Table. For example, COLUMN0 - EQ SOME VALUE. Type: string (or Expression with resultType string). - :type rfc_table_options: any - :param batch_size: Specifies the maximum number of rows that will be retrieved at a time when - retrieving data from SAP Table. Type: integer (or Expression with resultType integer). - :type batch_size: any - :param custom_rfc_read_table_function_module: Specifies the custom RFC function module that - will be used to read data from SAP Table. Type: string (or Expression with resultType string). - :type custom_rfc_read_table_function_module: any - :param sap_data_column_delimiter: The single character that will be used as delimiter passed to - SAP RFC as well as splitting the output data retrieved. Type: string (or Expression with - resultType string). - :type sap_data_column_delimiter: any - :param partition_option: The partition mechanism that will be used for SAP table read in - parallel. Possible values include: "None", "PartitionOnInt", "PartitionOnCalendarYear", - "PartitionOnCalendarMonth", "PartitionOnCalendarDate", "PartitionOnTime". - :type partition_option: str or - ~azure.synapse.artifacts.v2020_12_01.models.SapTablePartitionOption - :param partition_settings: The settings that will be leveraged for SAP table source - partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.SapTablePartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'row_count': {'key': 'rowCount', 'type': 'object'}, - 'row_skips': {'key': 'rowSkips', 'type': 'object'}, - 'rfc_table_fields': {'key': 'rfcTableFields', 'type': 'object'}, - 'rfc_table_options': {'key': 'rfcTableOptions', 'type': 'object'}, - 'batch_size': {'key': 'batchSize', 'type': 'object'}, - 'custom_rfc_read_table_function_module': {'key': 'customRfcReadTableFunctionModule', 'type': 'object'}, - 'sap_data_column_delimiter': {'key': 'sapDataColumnDelimiter', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'SapTablePartitionSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(SapTableSource, self).__init__(**kwargs) - self.type = 'SapTableSource' # type: str - self.row_count = kwargs.get('row_count', None) - self.row_skips = kwargs.get('row_skips', None) - self.rfc_table_fields = kwargs.get('rfc_table_fields', None) - self.rfc_table_options = kwargs.get('rfc_table_options', None) - self.batch_size = kwargs.get('batch_size', None) - self.custom_rfc_read_table_function_module = kwargs.get('custom_rfc_read_table_function_module', None) - self.sap_data_column_delimiter = kwargs.get('sap_data_column_delimiter', None) - self.partition_option = kwargs.get('partition_option', None) - self.partition_settings = kwargs.get('partition_settings', None) - - -class ScheduleTrigger(MultiplePipelineTrigger): - """Trigger that creates pipeline runs periodically, on schedule. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Trigger type.Constant filled by server. - :type type: str - :param description: Trigger description. - :type description: str - :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are - called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRuntimeState - :param annotations: List of tags that can be used for describing the trigger. - :type annotations: list[any] - :param pipelines: Pipelines that need to be started. - :type pipelines: list[~azure.synapse.artifacts.v2020_12_01.models.TriggerPipelineReference] - :param recurrence: Required. Recurrence schedule configuration. - :type recurrence: ~azure.synapse.artifacts.v2020_12_01.models.ScheduleTriggerRecurrence - """ - - _validation = { - 'type': {'required': True}, - 'runtime_state': {'readonly': True}, - 'recurrence': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, - 'recurrence': {'key': 'typeProperties.recurrence', 'type': 'ScheduleTriggerRecurrence'}, - } - - def __init__( - self, - **kwargs - ): - super(ScheduleTrigger, self).__init__(**kwargs) - self.type = 'ScheduleTrigger' # type: str - self.recurrence = kwargs['recurrence'] - - -class ScheduleTriggerRecurrence(msrest.serialization.Model): - """The workflow trigger recurrence. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param frequency: The frequency. Possible values include: "NotSpecified", "Minute", "Hour", - "Day", "Week", "Month", "Year". - :type frequency: str or ~azure.synapse.artifacts.v2020_12_01.models.RecurrenceFrequency - :param interval: The interval. - :type interval: int - :param start_time: The start time. - :type start_time: ~datetime.datetime - :param end_time: The end time. - :type end_time: ~datetime.datetime - :param time_zone: The time zone. - :type time_zone: str - :param schedule: The recurrence schedule. - :type schedule: ~azure.synapse.artifacts.v2020_12_01.models.RecurrenceSchedule - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'frequency': {'key': 'frequency', 'type': 'str'}, - 'interval': {'key': 'interval', 'type': 'int'}, - 'start_time': {'key': 'startTime', 'type': 'iso-8601'}, - 'end_time': {'key': 'endTime', 'type': 'iso-8601'}, - 'time_zone': {'key': 'timeZone', 'type': 'str'}, - 'schedule': {'key': 'schedule', 'type': 'RecurrenceSchedule'}, - } - - def __init__( - self, - **kwargs - ): - super(ScheduleTriggerRecurrence, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.frequency = kwargs.get('frequency', None) - self.interval = kwargs.get('interval', None) - self.start_time = kwargs.get('start_time', None) - self.end_time = kwargs.get('end_time', None) - self.time_zone = kwargs.get('time_zone', None) - self.schedule = kwargs.get('schedule', None) - - -class ScriptAction(msrest.serialization.Model): - """Custom script action to run on HDI ondemand cluster once it's up. - - All required parameters must be populated in order to send to Azure. - - :param name: Required. The user provided name of the script action. - :type name: str - :param uri: Required. The URI for the script action. - :type uri: str - :param roles: Required. The node types on which the script action should be executed. Possible - values include: "Headnode", "Workernode", "Zookeeper". - :type roles: str or ~azure.synapse.artifacts.v2020_12_01.models.HdiNodeTypes - :param parameters: The parameters for the script action. - :type parameters: str - """ - - _validation = { - 'name': {'required': True}, - 'uri': {'required': True}, - 'roles': {'required': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'uri': {'key': 'uri', 'type': 'str'}, - 'roles': {'key': 'roles', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(ScriptAction, self).__init__(**kwargs) - self.name = kwargs['name'] - self.uri = kwargs['uri'] - self.roles = kwargs['roles'] - self.parameters = kwargs.get('parameters', None) - - -class SecureString(SecretBase): - """Azure Synapse secure string definition. The string value will be masked with asterisks '*' during Get or List API calls. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Type of the secret.Constant filled by server. - :type type: str - :param value: Required. Value of secure string. - :type value: str - """ - - _validation = { - 'type': {'required': True}, - 'value': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'value': {'key': 'value', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SecureString, self).__init__(**kwargs) - self.type = 'SecureString' # type: str - self.value = kwargs['value'] - - -class SelfDependencyTumblingWindowTriggerReference(DependencyReference): - """Self referenced tumbling window trigger dependency. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. The type of dependency reference.Constant filled by server. - :type type: str - :param offset: Required. Timespan applied to the start time of a tumbling window when - evaluating dependency. - :type offset: str - :param size: The size of the window when evaluating the dependency. If undefined the frequency - of the tumbling window will be used. - :type size: str - """ - - _validation = { - 'type': {'required': True}, - 'offset': {'required': True, 'max_length': 15, 'min_length': 8, 'pattern': r'-((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))'}, - 'size': {'max_length': 15, 'min_length': 8, 'pattern': r'((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))'}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'offset': {'key': 'offset', 'type': 'str'}, - 'size': {'key': 'size', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SelfDependencyTumblingWindowTriggerReference, self).__init__(**kwargs) - self.type = 'SelfDependencyTumblingWindowTriggerReference' # type: str - self.offset = kwargs['offset'] - self.size = kwargs.get('size', None) - - -class SelfHostedIntegrationRuntime(IntegrationRuntime): - """Self-hosted integration runtime. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of integration runtime.Constant filled by server. Possible values - include: "Managed", "SelfHosted". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeType - :param description: Integration runtime description. - :type description: str - :param linked_info: Linked integration runtime type from data factory. - :type linked_info: ~azure.synapse.artifacts.v2020_12_01.models.LinkedIntegrationRuntimeType - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'linked_info': {'key': 'typeProperties.linkedInfo', 'type': 'LinkedIntegrationRuntimeType'}, - } - - def __init__( - self, - **kwargs - ): - super(SelfHostedIntegrationRuntime, self).__init__(**kwargs) - self.type = 'SelfHosted' # type: str - self.linked_info = kwargs.get('linked_info', None) - - -class ServiceNowLinkedService(LinkedService): - """ServiceNow server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param endpoint: Required. The endpoint of the ServiceNow server. (i.e. - :code:``.service-now.com). - :type endpoint: any - :param authentication_type: Required. The authentication type to use. Possible values include: - "Basic", "OAuth2". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.ServiceNowAuthenticationType - :param username: The user name used to connect to the ServiceNow server for Basic and OAuth2 - authentication. - :type username: any - :param password: The password corresponding to the user name for Basic and OAuth2 - authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param client_id: The client id for OAuth2 authentication. - :type client_id: any - :param client_secret: The client secret for OAuth2 authentication. - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'endpoint': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ServiceNowLinkedService, self).__init__(**kwargs) - self.type = 'ServiceNow' # type: str - self.endpoint = kwargs['endpoint'] - self.authentication_type = kwargs['authentication_type'] - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.client_id = kwargs.get('client_id', None) - self.client_secret = kwargs.get('client_secret', None) - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.use_host_verification = kwargs.get('use_host_verification', None) - self.use_peer_verification = kwargs.get('use_peer_verification', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class ServiceNowObjectDataset(Dataset): - """ServiceNow server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ServiceNowObjectDataset, self).__init__(**kwargs) - self.type = 'ServiceNowObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class ServiceNowSource(TabularSource): - """A copy activity ServiceNow server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ServiceNowSource, self).__init__(**kwargs) - self.type = 'ServiceNowSource' # type: str - self.query = kwargs.get('query', None) - - -class SetVariableActivity(ControlActivity): - """Set value for a Variable. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param variable_name: Name of the variable whose value needs to be set. - :type variable_name: str - :param value: Value to be set. Could be a static value or Expression. - :type value: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'variable_name': {'key': 'typeProperties.variableName', 'type': 'str'}, - 'value': {'key': 'typeProperties.value', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SetVariableActivity, self).__init__(**kwargs) - self.type = 'SetVariable' # type: str - self.variable_name = kwargs.get('variable_name', None) - self.value = kwargs.get('value', None) - - -class SftpLocation(DatasetLocation): - """The location of SFTP dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SftpLocation, self).__init__(**kwargs) - self.type = 'SftpLocation' # type: str - - -class SftpReadSettings(StoreReadSettings): - """Sftp read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: Sftp wildcardFolderPath. Type: string (or Expression with - resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: Sftp wildcardFileName. Type: string (or Expression with resultType - string). - :type wildcard_file_name: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - 'modified_datetime_start': {'key': 'modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'modifiedDatetimeEnd', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SftpReadSettings, self).__init__(**kwargs) - self.type = 'SftpReadSettings' # type: str - self.recursive = kwargs.get('recursive', None) - self.wildcard_folder_path = kwargs.get('wildcard_folder_path', None) - self.wildcard_file_name = kwargs.get('wildcard_file_name', None) - self.enable_partition_discovery = kwargs.get('enable_partition_discovery', None) - self.partition_root_path = kwargs.get('partition_root_path', None) - self.file_list_path = kwargs.get('file_list_path', None) - self.delete_files_after_completion = kwargs.get('delete_files_after_completion', None) - self.modified_datetime_start = kwargs.get('modified_datetime_start', None) - self.modified_datetime_end = kwargs.get('modified_datetime_end', None) - - -class SftpServerLinkedService(LinkedService): - """A linked service for an SSH File Transfer Protocol (SFTP) server. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The SFTP server host name. Type: string (or Expression with resultType - string). - :type host: any - :param port: The TCP port number that the SFTP server uses to listen for client connections. - Default value is 22. Type: integer (or Expression with resultType integer), minimum: 0. - :type port: any - :param authentication_type: The authentication type to be used to connect to the FTP server. - Possible values include: "Basic", "SshPublicKey". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.SftpAuthenticationType - :param user_name: The username used to log on to the SFTP server. Type: string (or Expression - with resultType string). - :type user_name: any - :param password: Password to logon the SFTP server for Basic authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - :param private_key_path: The SSH private key file path for SshPublicKey authentication. Only - valid for on-premises copy. For on-premises copy with SshPublicKey authentication, either - PrivateKeyPath or PrivateKeyContent should be specified. SSH private key should be OpenSSH - format. Type: string (or Expression with resultType string). - :type private_key_path: any - :param private_key_content: Base64 encoded SSH private key content for SshPublicKey - authentication. For on-premises copy with SshPublicKey authentication, either PrivateKeyPath or - PrivateKeyContent should be specified. SSH private key should be OpenSSH format. - :type private_key_content: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param pass_phrase: The password to decrypt the SSH private key if the SSH private key is - encrypted. - :type pass_phrase: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param skip_host_key_validation: If true, skip the SSH host key validation. Default value is - false. Type: boolean (or Expression with resultType boolean). - :type skip_host_key_validation: any - :param host_key_fingerprint: The host key finger-print of the SFTP server. When - SkipHostKeyValidation is false, HostKeyFingerprint should be specified. Type: string (or - Expression with resultType string). - :type host_key_fingerprint: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - 'private_key_path': {'key': 'typeProperties.privateKeyPath', 'type': 'object'}, - 'private_key_content': {'key': 'typeProperties.privateKeyContent', 'type': 'SecretBase'}, - 'pass_phrase': {'key': 'typeProperties.passPhrase', 'type': 'SecretBase'}, - 'skip_host_key_validation': {'key': 'typeProperties.skipHostKeyValidation', 'type': 'object'}, - 'host_key_fingerprint': {'key': 'typeProperties.hostKeyFingerprint', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SftpServerLinkedService, self).__init__(**kwargs) - self.type = 'Sftp' # type: str - self.host = kwargs['host'] - self.port = kwargs.get('port', None) - self.authentication_type = kwargs.get('authentication_type', None) - self.user_name = kwargs.get('user_name', None) - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - self.private_key_path = kwargs.get('private_key_path', None) - self.private_key_content = kwargs.get('private_key_content', None) - self.pass_phrase = kwargs.get('pass_phrase', None) - self.skip_host_key_validation = kwargs.get('skip_host_key_validation', None) - self.host_key_fingerprint = kwargs.get('host_key_fingerprint', None) - - -class SftpWriteSettings(StoreWriteSettings): - """Sftp write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - :param operation_timeout: Specifies the timeout for writing each chunk to SFTP server. Default - value: 01:00:00 (one hour). Type: string (or Expression with resultType string). - :type operation_timeout: any - :param use_temp_file_rename: Upload to temporary file(s) and rename. Disable this option if - your SFTP server doesn't support rename operation. Type: boolean (or Expression with resultType - boolean). - :type use_temp_file_rename: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - 'operation_timeout': {'key': 'operationTimeout', 'type': 'object'}, - 'use_temp_file_rename': {'key': 'useTempFileRename', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SftpWriteSettings, self).__init__(**kwargs) - self.type = 'SftpWriteSettings' # type: str - self.operation_timeout = kwargs.get('operation_timeout', None) - self.use_temp_file_rename = kwargs.get('use_temp_file_rename', None) - - -class SharePointOnlineListLinkedService(LinkedService): - """SharePoint Online List linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param site_url: Required. The URL of the SharePoint Online site. For example, - https://contoso.sharepoint.com/sites/siteName. Type: string (or Expression with resultType - string). - :type site_url: any - :param tenant_id: Required. The tenant ID under which your application resides. You can find it - from Azure portal Active Directory overview page. Type: string (or Expression with resultType - string). - :type tenant_id: any - :param service_principal_id: Required. The application (client) ID of your application - registered in Azure Active Directory. Make sure to grant SharePoint site permission to this - application. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: Required. The client secret of your application registered in - Azure Active Directory. Type: string (or Expression with resultType string). - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'site_url': {'required': True}, - 'tenant_id': {'required': True}, - 'service_principal_id': {'required': True}, - 'service_principal_key': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'site_url': {'key': 'typeProperties.siteUrl', 'type': 'object'}, - 'tenant_id': {'key': 'typeProperties.tenantId', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SharePointOnlineListLinkedService, self).__init__(**kwargs) - self.type = 'SharePointOnlineList' # type: str - self.site_url = kwargs['site_url'] - self.tenant_id = kwargs['tenant_id'] - self.service_principal_id = kwargs['service_principal_id'] - self.service_principal_key = kwargs['service_principal_key'] - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class SharePointOnlineListResourceDataset(Dataset): - """The sharepoint online list resource dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param list_name: The name of the SharePoint Online list. Type: string (or Expression with - resultType string). - :type list_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'list_name': {'key': 'typeProperties.listName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SharePointOnlineListResourceDataset, self).__init__(**kwargs) - self.type = 'SharePointOnlineListResource' # type: str - self.list_name = kwargs.get('list_name', None) - - -class SharePointOnlineListSource(CopySource): - """A copy activity source for sharePoint online list source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: The OData query to filter the data in SharePoint Online list. For example, - "$top=1". Type: string (or Expression with resultType string). - :type query: any - :param http_request_timeout: The wait time to get a response from SharePoint Online. Default - value is 5 minutes (00:05:00). Type: string (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SharePointOnlineListSource, self).__init__(**kwargs) - self.type = 'SharePointOnlineListSource' # type: str - self.query = kwargs.get('query', None) - self.http_request_timeout = kwargs.get('http_request_timeout', None) - - -class ShopifyLinkedService(LinkedService): - """Shopify Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The endpoint of the Shopify server. (i.e. mystore.myshopify.com). - :type host: any - :param access_token: The API access token that can be used to access Shopify’s data. The token - won't expire if it is offline mode. - :type access_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ShopifyLinkedService, self).__init__(**kwargs) - self.type = 'Shopify' # type: str - self.host = kwargs['host'] - self.access_token = kwargs.get('access_token', None) - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.use_host_verification = kwargs.get('use_host_verification', None) - self.use_peer_verification = kwargs.get('use_peer_verification', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class ShopifyObjectDataset(Dataset): - """Shopify Service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ShopifyObjectDataset, self).__init__(**kwargs) - self.type = 'ShopifyObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class ShopifySource(TabularSource): - """A copy activity Shopify Service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ShopifySource, self).__init__(**kwargs) - self.type = 'ShopifySource' # type: str - self.query = kwargs.get('query', None) - - -class SkipErrorFile(msrest.serialization.Model): - """Skip error file. - - :param file_missing: Skip if file is deleted by other client during copy. Default is true. - Type: boolean (or Expression with resultType boolean). - :type file_missing: any - :param data_inconsistency: Skip if source/sink file changed by other concurrent write. Default - is false. Type: boolean (or Expression with resultType boolean). - :type data_inconsistency: any - """ - - _attribute_map = { - 'file_missing': {'key': 'fileMissing', 'type': 'object'}, - 'data_inconsistency': {'key': 'dataInconsistency', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SkipErrorFile, self).__init__(**kwargs) - self.file_missing = kwargs.get('file_missing', None) - self.data_inconsistency = kwargs.get('data_inconsistency', None) - - -class Sku(msrest.serialization.Model): - """SQL pool SKU. - - :param tier: The service tier. - :type tier: str - :param name: The SKU name. - :type name: str - :param capacity: If the SKU supports scale out/in then the capacity integer should be included. - If scale out/in is not possible for the resource this may be omitted. - :type capacity: int - """ - - _attribute_map = { - 'tier': {'key': 'tier', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'capacity': {'key': 'capacity', 'type': 'int'}, - } - - def __init__( - self, - **kwargs - ): - super(Sku, self).__init__(**kwargs) - self.tier = kwargs.get('tier', None) - self.name = kwargs.get('name', None) - self.capacity = kwargs.get('capacity', None) - - -class SnowflakeDataset(Dataset): - """The snowflake dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param schema_type_properties_schema: The schema name of the Snowflake database. Type: string - (or Expression with resultType string). - :type schema_type_properties_schema: any - :param table: The table name of the Snowflake database. Type: string (or Expression with - resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SnowflakeDataset, self).__init__(**kwargs) - self.type = 'SnowflakeTable' # type: str - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - self.table = kwargs.get('table', None) - - -class SnowflakeExportCopyCommand(ExportSettings): - """Snowflake export command settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The export setting type.Constant filled by server. - :type type: str - :param additional_copy_options: Additional copy options directly passed to snowflake Copy - Command. Type: key value pairs (value should be string type) (or Expression with resultType - object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": - "'HH24:MI:SS.FF'" }. - :type additional_copy_options: dict[str, any] - :param additional_format_options: Additional format options directly passed to snowflake Copy - Command. Type: key value pairs (value should be string type) (or Expression with resultType - object). Example: "additionalFormatOptions": { "OVERWRITE": "TRUE", "MAX_FILE_SIZE": "'FALSE'" - }. - :type additional_format_options: dict[str, any] - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'additional_copy_options': {'key': 'additionalCopyOptions', 'type': '{object}'}, - 'additional_format_options': {'key': 'additionalFormatOptions', 'type': '{object}'}, - } - - def __init__( - self, - **kwargs - ): - super(SnowflakeExportCopyCommand, self).__init__(**kwargs) - self.type = 'SnowflakeExportCopyCommand' # type: str - self.additional_copy_options = kwargs.get('additional_copy_options', None) - self.additional_format_options = kwargs.get('additional_format_options', None) - - -class SnowflakeImportCopyCommand(ImportSettings): - """Snowflake import command settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The import setting type.Constant filled by server. - :type type: str - :param additional_copy_options: Additional copy options directly passed to snowflake Copy - Command. Type: key value pairs (value should be string type) (or Expression with resultType - object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": - "'HH24:MI:SS.FF'" }. - :type additional_copy_options: dict[str, any] - :param additional_format_options: Additional format options directly passed to snowflake Copy - Command. Type: key value pairs (value should be string type) (or Expression with resultType - object). Example: "additionalFormatOptions": { "FORCE": "TRUE", "LOAD_UNCERTAIN_FILES": - "'FALSE'" }. - :type additional_format_options: dict[str, any] - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'additional_copy_options': {'key': 'additionalCopyOptions', 'type': '{object}'}, - 'additional_format_options': {'key': 'additionalFormatOptions', 'type': '{object}'}, - } - - def __init__( - self, - **kwargs - ): - super(SnowflakeImportCopyCommand, self).__init__(**kwargs) - self.type = 'SnowflakeImportCopyCommand' # type: str - self.additional_copy_options = kwargs.get('additional_copy_options', None) - self.additional_format_options = kwargs.get('additional_format_options', None) - - -class SnowflakeLinkedService(LinkedService): - """Snowflake linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The connection string of snowflake. Type: string, - SecureString. - :type connection_string: any - :param password: The Azure key vault secret reference of password in connection string. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SnowflakeLinkedService, self).__init__(**kwargs) - self.type = 'Snowflake' # type: str - self.connection_string = kwargs['connection_string'] - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class SnowflakeSink(CopySink): - """A copy activity snowflake sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType - string). - :type pre_copy_script: any - :param import_settings: Snowflake import settings. - :type import_settings: ~azure.synapse.artifacts.v2020_12_01.models.SnowflakeImportCopyCommand - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - 'import_settings': {'key': 'importSettings', 'type': 'SnowflakeImportCopyCommand'}, - } - - def __init__( - self, - **kwargs - ): - super(SnowflakeSink, self).__init__(**kwargs) - self.type = 'SnowflakeSink' # type: str - self.pre_copy_script = kwargs.get('pre_copy_script', None) - self.import_settings = kwargs.get('import_settings', None) - - -class SnowflakeSource(CopySource): - """A copy activity snowflake source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: Snowflake Sql query. Type: string (or Expression with resultType string). - :type query: any - :param export_settings: Snowflake export settings. - :type export_settings: ~azure.synapse.artifacts.v2020_12_01.models.SnowflakeExportCopyCommand - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'export_settings': {'key': 'exportSettings', 'type': 'SnowflakeExportCopyCommand'}, - } - - def __init__( - self, - **kwargs - ): - super(SnowflakeSource, self).__init__(**kwargs) - self.type = 'SnowflakeSource' # type: str - self.query = kwargs.get('query', None) - self.export_settings = kwargs.get('export_settings', None) - - -class SparkBatchJob(msrest.serialization.Model): - """SparkBatchJob. - - All required parameters must be populated in order to send to Azure. - - :param livy_info: - :type livy_info: ~azure.synapse.artifacts.v2020_12_01.models.SparkBatchJobState - :param name: The batch name. - :type name: str - :param workspace_name: The workspace name. - :type workspace_name: str - :param spark_pool_name: The Spark pool name. - :type spark_pool_name: str - :param submitter_name: The submitter name. - :type submitter_name: str - :param submitter_id: The submitter identifier. - :type submitter_id: str - :param artifact_id: The artifact identifier. - :type artifact_id: str - :param job_type: The job type. Possible values include: "SparkBatch", "SparkSession". - :type job_type: str or ~azure.synapse.artifacts.v2020_12_01.models.SparkJobType - :param result: The Spark batch job result. Possible values include: "Uncertain", "Succeeded", - "Failed", "Cancelled". - :type result: str or ~azure.synapse.artifacts.v2020_12_01.models.SparkBatchJobResultType - :param scheduler: The scheduler information. - :type scheduler: ~azure.synapse.artifacts.v2020_12_01.models.SparkScheduler - :param plugin: The plugin information. - :type plugin: ~azure.synapse.artifacts.v2020_12_01.models.SparkServicePlugin - :param errors: The error information. - :type errors: list[~azure.synapse.artifacts.v2020_12_01.models.SparkServiceError] - :param tags: A set of tags. The tags. - :type tags: dict[str, str] - :param id: Required. The session Id. - :type id: int - :param app_id: The application id of this session. - :type app_id: str - :param app_info: The detailed application info. - :type app_info: dict[str, str] - :param state: The batch state. - :type state: str - :param log_lines: The log lines. - :type log_lines: list[str] - """ - - _validation = { - 'id': {'required': True}, - } - - _attribute_map = { - 'livy_info': {'key': 'livyInfo', 'type': 'SparkBatchJobState'}, - 'name': {'key': 'name', 'type': 'str'}, - 'workspace_name': {'key': 'workspaceName', 'type': 'str'}, - 'spark_pool_name': {'key': 'sparkPoolName', 'type': 'str'}, - 'submitter_name': {'key': 'submitterName', 'type': 'str'}, - 'submitter_id': {'key': 'submitterId', 'type': 'str'}, - 'artifact_id': {'key': 'artifactId', 'type': 'str'}, - 'job_type': {'key': 'jobType', 'type': 'str'}, - 'result': {'key': 'result', 'type': 'str'}, - 'scheduler': {'key': 'schedulerInfo', 'type': 'SparkScheduler'}, - 'plugin': {'key': 'pluginInfo', 'type': 'SparkServicePlugin'}, - 'errors': {'key': 'errorInfo', 'type': '[SparkServiceError]'}, - 'tags': {'key': 'tags', 'type': '{str}'}, - 'id': {'key': 'id', 'type': 'int'}, - 'app_id': {'key': 'appId', 'type': 'str'}, - 'app_info': {'key': 'appInfo', 'type': '{str}'}, - 'state': {'key': 'state', 'type': 'str'}, - 'log_lines': {'key': 'log', 'type': '[str]'}, - } - - def __init__( - self, - **kwargs - ): - super(SparkBatchJob, self).__init__(**kwargs) - self.livy_info = kwargs.get('livy_info', None) - self.name = kwargs.get('name', None) - self.workspace_name = kwargs.get('workspace_name', None) - self.spark_pool_name = kwargs.get('spark_pool_name', None) - self.submitter_name = kwargs.get('submitter_name', None) - self.submitter_id = kwargs.get('submitter_id', None) - self.artifact_id = kwargs.get('artifact_id', None) - self.job_type = kwargs.get('job_type', None) - self.result = kwargs.get('result', None) - self.scheduler = kwargs.get('scheduler', None) - self.plugin = kwargs.get('plugin', None) - self.errors = kwargs.get('errors', None) - self.tags = kwargs.get('tags', None) - self.id = kwargs['id'] - self.app_id = kwargs.get('app_id', None) - self.app_info = kwargs.get('app_info', None) - self.state = kwargs.get('state', None) - self.log_lines = kwargs.get('log_lines', None) - - -class SparkBatchJobState(msrest.serialization.Model): - """SparkBatchJobState. - - :param not_started_at: the time that at which "not_started" livy state was first seen. - :type not_started_at: ~datetime.datetime - :param starting_at: the time that at which "starting" livy state was first seen. - :type starting_at: ~datetime.datetime - :param running_at: the time that at which "running" livy state was first seen. - :type running_at: ~datetime.datetime - :param dead_at: time that at which "dead" livy state was first seen. - :type dead_at: ~datetime.datetime - :param success_at: the time that at which "success" livy state was first seen. - :type success_at: ~datetime.datetime - :param terminated_at: the time that at which "killed" livy state was first seen. - :type terminated_at: ~datetime.datetime - :param recovering_at: the time that at which "recovering" livy state was first seen. - :type recovering_at: ~datetime.datetime - :param current_state: the Spark job state. - :type current_state: str - :param job_creation_request: - :type job_creation_request: ~azure.synapse.artifacts.v2020_12_01.models.SparkRequest - """ - - _attribute_map = { - 'not_started_at': {'key': 'notStartedAt', 'type': 'iso-8601'}, - 'starting_at': {'key': 'startingAt', 'type': 'iso-8601'}, - 'running_at': {'key': 'runningAt', 'type': 'iso-8601'}, - 'dead_at': {'key': 'deadAt', 'type': 'iso-8601'}, - 'success_at': {'key': 'successAt', 'type': 'iso-8601'}, - 'terminated_at': {'key': 'killedAt', 'type': 'iso-8601'}, - 'recovering_at': {'key': 'recoveringAt', 'type': 'iso-8601'}, - 'current_state': {'key': 'currentState', 'type': 'str'}, - 'job_creation_request': {'key': 'jobCreationRequest', 'type': 'SparkRequest'}, - } - - def __init__( - self, - **kwargs - ): - super(SparkBatchJobState, self).__init__(**kwargs) - self.not_started_at = kwargs.get('not_started_at', None) - self.starting_at = kwargs.get('starting_at', None) - self.running_at = kwargs.get('running_at', None) - self.dead_at = kwargs.get('dead_at', None) - self.success_at = kwargs.get('success_at', None) - self.terminated_at = kwargs.get('terminated_at', None) - self.recovering_at = kwargs.get('recovering_at', None) - self.current_state = kwargs.get('current_state', None) - self.job_creation_request = kwargs.get('job_creation_request', None) - - -class SparkJobDefinition(msrest.serialization.Model): - """Spark job definition. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param description: The description of the Spark job definition. - :type description: str - :param target_big_data_pool: Required. Big data pool reference. - :type target_big_data_pool: ~azure.synapse.artifacts.v2020_12_01.models.BigDataPoolReference - :param required_spark_version: The required Spark version of the application. - :type required_spark_version: str - :param language: The language of the Spark application. - :type language: str - :param job_properties: Required. The properties of the Spark job. - :type job_properties: ~azure.synapse.artifacts.v2020_12_01.models.SparkJobProperties - """ - - _validation = { - 'target_big_data_pool': {'required': True}, - 'job_properties': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'description', 'type': 'str'}, - 'target_big_data_pool': {'key': 'targetBigDataPool', 'type': 'BigDataPoolReference'}, - 'required_spark_version': {'key': 'requiredSparkVersion', 'type': 'str'}, - 'language': {'key': 'language', 'type': 'str'}, - 'job_properties': {'key': 'jobProperties', 'type': 'SparkJobProperties'}, - } - - def __init__( - self, - **kwargs - ): - super(SparkJobDefinition, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.description = kwargs.get('description', None) - self.target_big_data_pool = kwargs['target_big_data_pool'] - self.required_spark_version = kwargs.get('required_spark_version', None) - self.language = kwargs.get('language', None) - self.job_properties = kwargs['job_properties'] - - -class SparkJobDefinitionResource(SubResource): - """Spark job definition resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Properties of spark job definition. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.SparkJobDefinition - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'SparkJobDefinition'}, - } - - def __init__( - self, - **kwargs - ): - super(SparkJobDefinitionResource, self).__init__(**kwargs) - self.properties = kwargs['properties'] - - -class SparkJobDefinitionsListResponse(msrest.serialization.Model): - """A list of spark job definitions resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of spark job definitions. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.SparkJobDefinitionResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[SparkJobDefinitionResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SparkJobDefinitionsListResponse, self).__init__(**kwargs) - self.value = kwargs['value'] - self.next_link = kwargs.get('next_link', None) - - -class SparkJobProperties(msrest.serialization.Model): - """The properties of the Spark job. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: The name of the job. - :type name: str - :param file: Required. File containing the application to execute. - :type file: str - :param class_name: Main class for Java/Scala application. - :type class_name: str - :param conf: Spark configuration properties. - :type conf: any - :param args: Command line arguments for the application. - :type args: list[str] - :param jars: Jars to be used in this job. - :type jars: list[str] - :param files: files to be used in this job. - :type files: list[str] - :param archives: Archives to be used in this job. - :type archives: list[str] - :param driver_memory: Required. Amount of memory to use for the driver process. - :type driver_memory: str - :param driver_cores: Required. Number of cores to use for the driver. - :type driver_cores: int - :param executor_memory: Required. Amount of memory to use per executor process. - :type executor_memory: str - :param executor_cores: Required. Number of cores to use for each executor. - :type executor_cores: int - :param num_executors: Required. Number of executors to launch for this job. - :type num_executors: int - """ - - _validation = { - 'file': {'required': True}, - 'driver_memory': {'required': True}, - 'driver_cores': {'required': True}, - 'executor_memory': {'required': True}, - 'executor_cores': {'required': True}, - 'num_executors': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'file': {'key': 'file', 'type': 'str'}, - 'class_name': {'key': 'className', 'type': 'str'}, - 'conf': {'key': 'conf', 'type': 'object'}, - 'args': {'key': 'args', 'type': '[str]'}, - 'jars': {'key': 'jars', 'type': '[str]'}, - 'files': {'key': 'files', 'type': '[str]'}, - 'archives': {'key': 'archives', 'type': '[str]'}, - 'driver_memory': {'key': 'driverMemory', 'type': 'str'}, - 'driver_cores': {'key': 'driverCores', 'type': 'int'}, - 'executor_memory': {'key': 'executorMemory', 'type': 'str'}, - 'executor_cores': {'key': 'executorCores', 'type': 'int'}, - 'num_executors': {'key': 'numExecutors', 'type': 'int'}, - } - - def __init__( - self, - **kwargs - ): - super(SparkJobProperties, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.name = kwargs.get('name', None) - self.file = kwargs['file'] - self.class_name = kwargs.get('class_name', None) - self.conf = kwargs.get('conf', None) - self.args = kwargs.get('args', None) - self.jars = kwargs.get('jars', None) - self.files = kwargs.get('files', None) - self.archives = kwargs.get('archives', None) - self.driver_memory = kwargs['driver_memory'] - self.driver_cores = kwargs['driver_cores'] - self.executor_memory = kwargs['executor_memory'] - self.executor_cores = kwargs['executor_cores'] - self.num_executors = kwargs['num_executors'] - - -class SparkLinkedService(LinkedService): - """Spark Server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. IP address or host name of the Spark server. - :type host: any - :param port: Required. The TCP port that the Spark server uses to listen for client - connections. - :type port: any - :param server_type: The type of Spark server. Possible values include: "SharkServer", - "SharkServer2", "SparkThriftServer". - :type server_type: str or ~azure.synapse.artifacts.v2020_12_01.models.SparkServerType - :param thrift_transport_protocol: The transport protocol to use in the Thrift layer. Possible - values include: "Binary", "SASL", "HTTP ". - :type thrift_transport_protocol: str or - ~azure.synapse.artifacts.v2020_12_01.models.SparkThriftTransportProtocol - :param authentication_type: Required. The authentication method used to access the Spark - server. Possible values include: "Anonymous", "Username", "UsernameAndPassword", - "WindowsAzureHDInsightService". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.SparkAuthenticationType - :param username: The user name that you use to access Spark Server. - :type username: any - :param password: The password corresponding to the user name that you provided in the Username - field. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param http_path: The partial URL corresponding to the Spark server. - :type http_path: any - :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The - default value is false. - :type enable_ssl: any - :param trusted_cert_path: The full path of the .pem file containing trusted CA certificates for - verifying the server when connecting over SSL. This property can only be set when using SSL on - self-hosted IR. The default value is the cacerts.pem file installed with the IR. - :type trusted_cert_path: any - :param use_system_trust_store: Specifies whether to use a CA certificate from the system trust - store or from a specified PEM file. The default value is false. - :type use_system_trust_store: any - :param allow_host_name_cn_mismatch: Specifies whether to require a CA-issued SSL certificate - name to match the host name of the server when connecting over SSL. The default value is false. - :type allow_host_name_cn_mismatch: any - :param allow_self_signed_server_cert: Specifies whether to allow self-signed certificates from - the server. The default value is false. - :type allow_self_signed_server_cert: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'port': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'server_type': {'key': 'typeProperties.serverType', 'type': 'str'}, - 'thrift_transport_protocol': {'key': 'typeProperties.thriftTransportProtocol', 'type': 'str'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'http_path': {'key': 'typeProperties.httpPath', 'type': 'object'}, - 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, - 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, - 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, - 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, - 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SparkLinkedService, self).__init__(**kwargs) - self.type = 'Spark' # type: str - self.host = kwargs['host'] - self.port = kwargs['port'] - self.server_type = kwargs.get('server_type', None) - self.thrift_transport_protocol = kwargs.get('thrift_transport_protocol', None) - self.authentication_type = kwargs['authentication_type'] - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.http_path = kwargs.get('http_path', None) - self.enable_ssl = kwargs.get('enable_ssl', None) - self.trusted_cert_path = kwargs.get('trusted_cert_path', None) - self.use_system_trust_store = kwargs.get('use_system_trust_store', None) - self.allow_host_name_cn_mismatch = kwargs.get('allow_host_name_cn_mismatch', None) - self.allow_self_signed_server_cert = kwargs.get('allow_self_signed_server_cert', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class SparkObjectDataset(Dataset): - """Spark Server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The table name of the Spark. Type: string (or Expression with resultType string). - :type table: any - :param schema_type_properties_schema: The schema name of the Spark. Type: string (or Expression - with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SparkObjectDataset, self).__init__(**kwargs) - self.type = 'SparkObject' # type: str - self.table_name = kwargs.get('table_name', None) - self.table = kwargs.get('table', None) - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - - -class SparkRequest(msrest.serialization.Model): - """SparkRequest. - - :param name: - :type name: str - :param file: - :type file: str - :param class_name: - :type class_name: str - :param arguments: - :type arguments: list[str] - :param jars: - :type jars: list[str] - :param python_files: - :type python_files: list[str] - :param files: - :type files: list[str] - :param archives: - :type archives: list[str] - :param configuration: Dictionary of :code:``. - :type configuration: dict[str, str] - :param driver_memory: - :type driver_memory: str - :param driver_cores: - :type driver_cores: int - :param executor_memory: - :type executor_memory: str - :param executor_cores: - :type executor_cores: int - :param executor_count: - :type executor_count: int - """ - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'file': {'key': 'file', 'type': 'str'}, - 'class_name': {'key': 'className', 'type': 'str'}, - 'arguments': {'key': 'args', 'type': '[str]'}, - 'jars': {'key': 'jars', 'type': '[str]'}, - 'python_files': {'key': 'pyFiles', 'type': '[str]'}, - 'files': {'key': 'files', 'type': '[str]'}, - 'archives': {'key': 'archives', 'type': '[str]'}, - 'configuration': {'key': 'conf', 'type': '{str}'}, - 'driver_memory': {'key': 'driverMemory', 'type': 'str'}, - 'driver_cores': {'key': 'driverCores', 'type': 'int'}, - 'executor_memory': {'key': 'executorMemory', 'type': 'str'}, - 'executor_cores': {'key': 'executorCores', 'type': 'int'}, - 'executor_count': {'key': 'numExecutors', 'type': 'int'}, - } - - def __init__( - self, - **kwargs - ): - super(SparkRequest, self).__init__(**kwargs) - self.name = kwargs.get('name', None) - self.file = kwargs.get('file', None) - self.class_name = kwargs.get('class_name', None) - self.arguments = kwargs.get('arguments', None) - self.jars = kwargs.get('jars', None) - self.python_files = kwargs.get('python_files', None) - self.files = kwargs.get('files', None) - self.archives = kwargs.get('archives', None) - self.configuration = kwargs.get('configuration', None) - self.driver_memory = kwargs.get('driver_memory', None) - self.driver_cores = kwargs.get('driver_cores', None) - self.executor_memory = kwargs.get('executor_memory', None) - self.executor_cores = kwargs.get('executor_cores', None) - self.executor_count = kwargs.get('executor_count', None) - - -class SparkScheduler(msrest.serialization.Model): - """SparkScheduler. - - :param submitted_at: - :type submitted_at: ~datetime.datetime - :param scheduled_at: - :type scheduled_at: ~datetime.datetime - :param ended_at: - :type ended_at: ~datetime.datetime - :param cancellation_requested_at: - :type cancellation_requested_at: ~datetime.datetime - :param current_state: Possible values include: "Queued", "Scheduled", "Ended". - :type current_state: str or ~azure.synapse.artifacts.v2020_12_01.models.SchedulerCurrentState - """ - - _attribute_map = { - 'submitted_at': {'key': 'submittedAt', 'type': 'iso-8601'}, - 'scheduled_at': {'key': 'scheduledAt', 'type': 'iso-8601'}, - 'ended_at': {'key': 'endedAt', 'type': 'iso-8601'}, - 'cancellation_requested_at': {'key': 'cancellationRequestedAt', 'type': 'iso-8601'}, - 'current_state': {'key': 'currentState', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SparkScheduler, self).__init__(**kwargs) - self.submitted_at = kwargs.get('submitted_at', None) - self.scheduled_at = kwargs.get('scheduled_at', None) - self.ended_at = kwargs.get('ended_at', None) - self.cancellation_requested_at = kwargs.get('cancellation_requested_at', None) - self.current_state = kwargs.get('current_state', None) - - -class SparkServiceError(msrest.serialization.Model): - """SparkServiceError. - - :param message: - :type message: str - :param error_code: - :type error_code: str - :param source: Possible values include: "System", "User", "Unknown", "Dependency". - :type source: str or ~azure.synapse.artifacts.v2020_12_01.models.SparkErrorSource - """ - - _attribute_map = { - 'message': {'key': 'message', 'type': 'str'}, - 'error_code': {'key': 'errorCode', 'type': 'str'}, - 'source': {'key': 'source', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SparkServiceError, self).__init__(**kwargs) - self.message = kwargs.get('message', None) - self.error_code = kwargs.get('error_code', None) - self.source = kwargs.get('source', None) - - -class SparkServicePlugin(msrest.serialization.Model): - """SparkServicePlugin. - - :param preparation_started_at: - :type preparation_started_at: ~datetime.datetime - :param resource_acquisition_started_at: - :type resource_acquisition_started_at: ~datetime.datetime - :param submission_started_at: - :type submission_started_at: ~datetime.datetime - :param monitoring_started_at: - :type monitoring_started_at: ~datetime.datetime - :param cleanup_started_at: - :type cleanup_started_at: ~datetime.datetime - :param current_state: Possible values include: "Preparation", "ResourceAcquisition", "Queued", - "Submission", "Monitoring", "Cleanup", "Ended". - :type current_state: str or ~azure.synapse.artifacts.v2020_12_01.models.PluginCurrentState - """ - - _attribute_map = { - 'preparation_started_at': {'key': 'preparationStartedAt', 'type': 'iso-8601'}, - 'resource_acquisition_started_at': {'key': 'resourceAcquisitionStartedAt', 'type': 'iso-8601'}, - 'submission_started_at': {'key': 'submissionStartedAt', 'type': 'iso-8601'}, - 'monitoring_started_at': {'key': 'monitoringStartedAt', 'type': 'iso-8601'}, - 'cleanup_started_at': {'key': 'cleanupStartedAt', 'type': 'iso-8601'}, - 'current_state': {'key': 'currentState', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SparkServicePlugin, self).__init__(**kwargs) - self.preparation_started_at = kwargs.get('preparation_started_at', None) - self.resource_acquisition_started_at = kwargs.get('resource_acquisition_started_at', None) - self.submission_started_at = kwargs.get('submission_started_at', None) - self.monitoring_started_at = kwargs.get('monitoring_started_at', None) - self.cleanup_started_at = kwargs.get('cleanup_started_at', None) - self.current_state = kwargs.get('current_state', None) - - -class SparkSource(TabularSource): - """A copy activity Spark Server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SparkSource, self).__init__(**kwargs) - self.type = 'SparkSource' # type: str - self.query = kwargs.get('query', None) - - -class SqlConnection(msrest.serialization.Model): - """The connection used to execute the SQL script. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The type of the connection. Possible values include: "SqlOnDemand", - "SqlPool". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.SqlConnectionType - :param name: Required. The identifier of the connection. - :type name: str - """ - - _validation = { - 'type': {'required': True}, - 'name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlConnection, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.type = kwargs['type'] - self.name = kwargs['name'] - - -class SqlDWSink(CopySink): - """A copy activity SQL Data Warehouse sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType - string). - :type pre_copy_script: any - :param allow_poly_base: Indicates to use PolyBase to copy data into SQL Data Warehouse when - applicable. Type: boolean (or Expression with resultType boolean). - :type allow_poly_base: any - :param poly_base_settings: Specifies PolyBase-related settings when allowPolyBase is true. - :type poly_base_settings: ~azure.synapse.artifacts.v2020_12_01.models.PolybaseSettings - :param allow_copy_command: Indicates to use Copy Command to copy data into SQL Data Warehouse. - Type: boolean (or Expression with resultType boolean). - :type allow_copy_command: any - :param copy_command_settings: Specifies Copy Command related settings when allowCopyCommand is - true. - :type copy_command_settings: ~azure.synapse.artifacts.v2020_12_01.models.DWCopyCommandSettings - :param table_option: The option to handle sink table, such as autoCreate. For now only - 'autoCreate' value is supported. Type: string (or Expression with resultType string). - :type table_option: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - 'allow_poly_base': {'key': 'allowPolyBase', 'type': 'object'}, - 'poly_base_settings': {'key': 'polyBaseSettings', 'type': 'PolybaseSettings'}, - 'allow_copy_command': {'key': 'allowCopyCommand', 'type': 'object'}, - 'copy_command_settings': {'key': 'copyCommandSettings', 'type': 'DWCopyCommandSettings'}, - 'table_option': {'key': 'tableOption', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlDWSink, self).__init__(**kwargs) - self.type = 'SqlDWSink' # type: str - self.pre_copy_script = kwargs.get('pre_copy_script', None) - self.allow_poly_base = kwargs.get('allow_poly_base', None) - self.poly_base_settings = kwargs.get('poly_base_settings', None) - self.allow_copy_command = kwargs.get('allow_copy_command', None) - self.copy_command_settings = kwargs.get('copy_command_settings', None) - self.table_option = kwargs.get('table_option', None) - - -class SqlDWSource(TabularSource): - """A copy activity SQL Data Warehouse source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param sql_reader_query: SQL Data Warehouse reader query. Type: string (or Expression with - resultType string). - :type sql_reader_query: any - :param sql_reader_stored_procedure_name: Name of the stored procedure for a SQL Data Warehouse - source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression - with resultType string). - :type sql_reader_stored_procedure_name: any - :param stored_procedure_parameters: Value and type setting for stored procedure parameters. - Example: "{Parameter1: {value: "1", type: "int"}}". Type: object (or Expression with resultType - object), itemType: StoredProcedureParameter. - :type stored_procedure_parameters: any - :param partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". - :type partition_option: any - :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.SqlPartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, - 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'object'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlDWSource, self).__init__(**kwargs) - self.type = 'SqlDWSource' # type: str - self.sql_reader_query = kwargs.get('sql_reader_query', None) - self.sql_reader_stored_procedure_name = kwargs.get('sql_reader_stored_procedure_name', None) - self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) - self.partition_option = kwargs.get('partition_option', None) - self.partition_settings = kwargs.get('partition_settings', None) - - -class SqlMISink(CopySink): - """A copy activity Azure SQL Managed Instance sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param sql_writer_stored_procedure_name: SQL writer stored procedure name. Type: string (or - Expression with resultType string). - :type sql_writer_stored_procedure_name: any - :param sql_writer_table_type: SQL writer table type. Type: string (or Expression with - resultType string). - :type sql_writer_table_type: any - :param pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType - string). - :type pre_copy_script: any - :param stored_procedure_parameters: SQL stored procedure parameters. - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - :param stored_procedure_table_type_parameter_name: The stored procedure parameter name of the - table type. Type: string (or Expression with resultType string). - :type stored_procedure_table_type_parameter_name: any - :param table_option: The option to handle sink table, such as autoCreate. For now only - 'autoCreate' value is supported. Type: string (or Expression with resultType string). - :type table_option: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'sql_writer_stored_procedure_name': {'key': 'sqlWriterStoredProcedureName', 'type': 'object'}, - 'sql_writer_table_type': {'key': 'sqlWriterTableType', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - 'stored_procedure_table_type_parameter_name': {'key': 'storedProcedureTableTypeParameterName', 'type': 'object'}, - 'table_option': {'key': 'tableOption', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlMISink, self).__init__(**kwargs) - self.type = 'SqlMISink' # type: str - self.sql_writer_stored_procedure_name = kwargs.get('sql_writer_stored_procedure_name', None) - self.sql_writer_table_type = kwargs.get('sql_writer_table_type', None) - self.pre_copy_script = kwargs.get('pre_copy_script', None) - self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) - self.stored_procedure_table_type_parameter_name = kwargs.get('stored_procedure_table_type_parameter_name', None) - self.table_option = kwargs.get('table_option', None) - - -class SqlMISource(TabularSource): - """A copy activity Azure SQL Managed Instance source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). - :type sql_reader_query: any - :param sql_reader_stored_procedure_name: Name of the stored procedure for a Azure SQL Managed - Instance source. This cannot be used at the same time as SqlReaderQuery. Type: string (or - Expression with resultType string). - :type sql_reader_stored_procedure_name: any - :param stored_procedure_parameters: Value and type setting for stored procedure parameters. - Example: "{Parameter1: {value: "1", type: "int"}}". - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - :param produce_additional_types: Which additional types to produce. - :type produce_additional_types: any - :param partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". - :type partition_option: any - :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.SqlPartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, - 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - 'produce_additional_types': {'key': 'produceAdditionalTypes', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'object'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlMISource, self).__init__(**kwargs) - self.type = 'SqlMISource' # type: str - self.sql_reader_query = kwargs.get('sql_reader_query', None) - self.sql_reader_stored_procedure_name = kwargs.get('sql_reader_stored_procedure_name', None) - self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) - self.produce_additional_types = kwargs.get('produce_additional_types', None) - self.partition_option = kwargs.get('partition_option', None) - self.partition_settings = kwargs.get('partition_settings', None) - - -class SqlPartitionSettings(msrest.serialization.Model): - """The settings that will be leveraged for Sql source partitioning. - - :param partition_column_name: The name of the column in integer or datetime type that will be - used for proceeding partitioning. If not specified, the primary key of the table is - auto-detected and used as the partition column. Type: string (or Expression with resultType - string). - :type partition_column_name: any - :param partition_upper_bound: The maximum value of the partition column for partition range - splitting. This value is used to decide the partition stride, not for filtering the rows in - table. All rows in the table or query result will be partitioned and copied. Type: string (or - Expression with resultType string). - :type partition_upper_bound: any - :param partition_lower_bound: The minimum value of the partition column for partition range - splitting. This value is used to decide the partition stride, not for filtering the rows in - table. All rows in the table or query result will be partitioned and copied. Type: string (or - Expression with resultType string). - :type partition_lower_bound: any - """ - - _attribute_map = { - 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, - 'partition_upper_bound': {'key': 'partitionUpperBound', 'type': 'object'}, - 'partition_lower_bound': {'key': 'partitionLowerBound', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlPartitionSettings, self).__init__(**kwargs) - self.partition_column_name = kwargs.get('partition_column_name', None) - self.partition_upper_bound = kwargs.get('partition_upper_bound', None) - self.partition_lower_bound = kwargs.get('partition_lower_bound', None) - - -class SqlPool(TrackedResource): - """A SQL Analytics pool. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :param tags: A set of tags. Resource tags. - :type tags: dict[str, str] - :param location: Required. The geo-location where the resource lives. - :type location: str - :param sku: SQL pool SKU. - :type sku: ~azure.synapse.artifacts.v2020_12_01.models.Sku - :param max_size_bytes: Maximum size in bytes. - :type max_size_bytes: long - :param collation: Collation mode. - :type collation: str - :param source_database_id: Source database to create from. - :type source_database_id: str - :param recoverable_database_id: Backup database to restore from. - :type recoverable_database_id: str - :param provisioning_state: Resource state. - :type provisioning_state: str - :param status: Resource status. - :type status: str - :param restore_point_in_time: Snapshot time to restore. - :type restore_point_in_time: str - :param create_mode: What is this?. - :type create_mode: str - :param creation_date: Date the SQL pool was created. - :type creation_date: ~datetime.datetime - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'location': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'tags': {'key': 'tags', 'type': '{str}'}, - 'location': {'key': 'location', 'type': 'str'}, - 'sku': {'key': 'sku', 'type': 'Sku'}, - 'max_size_bytes': {'key': 'properties.maxSizeBytes', 'type': 'long'}, - 'collation': {'key': 'properties.collation', 'type': 'str'}, - 'source_database_id': {'key': 'properties.sourceDatabaseId', 'type': 'str'}, - 'recoverable_database_id': {'key': 'properties.recoverableDatabaseId', 'type': 'str'}, - 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, - 'status': {'key': 'properties.status', 'type': 'str'}, - 'restore_point_in_time': {'key': 'properties.restorePointInTime', 'type': 'str'}, - 'create_mode': {'key': 'properties.createMode', 'type': 'str'}, - 'creation_date': {'key': 'properties.creationDate', 'type': 'iso-8601'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlPool, self).__init__(**kwargs) - self.sku = kwargs.get('sku', None) - self.max_size_bytes = kwargs.get('max_size_bytes', None) - self.collation = kwargs.get('collation', None) - self.source_database_id = kwargs.get('source_database_id', None) - self.recoverable_database_id = kwargs.get('recoverable_database_id', None) - self.provisioning_state = kwargs.get('provisioning_state', None) - self.status = kwargs.get('status', None) - self.restore_point_in_time = kwargs.get('restore_point_in_time', None) - self.create_mode = kwargs.get('create_mode', None) - self.creation_date = kwargs.get('creation_date', None) - - -class SqlPoolInfoListResult(msrest.serialization.Model): - """List of SQL pools. - - :param next_link: Link to the next page of results. - :type next_link: str - :param value: List of SQL pools. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.SqlPool] - """ - - _attribute_map = { - 'next_link': {'key': 'nextLink', 'type': 'str'}, - 'value': {'key': 'value', 'type': '[SqlPool]'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlPoolInfoListResult, self).__init__(**kwargs) - self.next_link = kwargs.get('next_link', None) - self.value = kwargs.get('value', None) - - -class SqlPoolReference(msrest.serialization.Model): - """SQL pool reference type. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. SQL pool reference type. Possible values include: "SqlPoolReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.SqlPoolReferenceType - :param reference_name: Required. Reference SQL pool name. - :type reference_name: str - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlPoolReference, self).__init__(**kwargs) - self.type = kwargs['type'] - self.reference_name = kwargs['reference_name'] - - -class SqlPoolStoredProcedureActivity(Activity): - """Execute SQL pool stored procedure activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param sql_pool: Required. SQL pool stored procedure reference. - :type sql_pool: ~azure.synapse.artifacts.v2020_12_01.models.SqlPoolReference - :param stored_procedure_name: Required. Stored procedure name. Type: string (or Expression with - resultType string). - :type stored_procedure_name: any - :param stored_procedure_parameters: Value and type setting for stored procedure parameters. - Example: "{Parameter1: {value: "1", type: "int"}}". - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'sql_pool': {'required': True}, - 'stored_procedure_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'sql_pool': {'key': 'sqlPool', 'type': 'SqlPoolReference'}, - 'stored_procedure_name': {'key': 'typeProperties.storedProcedureName', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'typeProperties.storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlPoolStoredProcedureActivity, self).__init__(**kwargs) - self.type = 'SqlPoolStoredProcedure' # type: str - self.sql_pool = kwargs['sql_pool'] - self.stored_procedure_name = kwargs['stored_procedure_name'] - self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) - - -class SqlScript(msrest.serialization.Model): - """SQL script. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param description: The description of the SQL script. - :type description: str - :param type: The type of the SQL script. Possible values include: "SqlQuery". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.SqlScriptType - :param content: Required. The content of the SQL script. - :type content: ~azure.synapse.artifacts.v2020_12_01.models.SqlScriptContent - """ - - _validation = { - 'content': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'description', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'content': {'key': 'content', 'type': 'SqlScriptContent'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlScript, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.description = kwargs.get('description', None) - self.type = kwargs.get('type', None) - self.content = kwargs['content'] - - -class SqlScriptContent(msrest.serialization.Model): - """The content of the SQL script. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param query: Required. SQL query to execute. - :type query: str - :param current_connection: Required. The connection used to execute the SQL script. - :type current_connection: ~azure.synapse.artifacts.v2020_12_01.models.SqlConnection - :param metadata: The metadata of the SQL script. - :type metadata: ~azure.synapse.artifacts.v2020_12_01.models.SqlScriptMetadata - """ - - _validation = { - 'query': {'required': True}, - 'current_connection': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'query': {'key': 'query', 'type': 'str'}, - 'current_connection': {'key': 'currentConnection', 'type': 'SqlConnection'}, - 'metadata': {'key': 'metadata', 'type': 'SqlScriptMetadata'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlScriptContent, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.query = kwargs['query'] - self.current_connection = kwargs['current_connection'] - self.metadata = kwargs.get('metadata', None) - - -class SqlScriptMetadata(msrest.serialization.Model): - """The metadata of the SQL script. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param language: The language of the SQL script. - :type language: str - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'language': {'key': 'language', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlScriptMetadata, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.language = kwargs.get('language', None) - - -class SqlScriptResource(msrest.serialization.Model): - """Sql Script resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource Id for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :param name: Required. The name of the resource. - :type name: str - :ivar type: The type of the resource. Ex- Microsoft.Compute/virtualMachines or - Microsoft.Storage/storageAccounts. - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Properties of sql script. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.SqlScript - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'required': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'SqlScript'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlScriptResource, self).__init__(**kwargs) - self.id = None - self.name = kwargs['name'] - self.type = None - self.etag = None - self.properties = kwargs['properties'] - - -class SqlScriptsListResponse(msrest.serialization.Model): - """A list of sql scripts resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of sql scripts. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.SqlScriptResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[SqlScriptResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlScriptsListResponse, self).__init__(**kwargs) - self.value = kwargs['value'] - self.next_link = kwargs.get('next_link', None) - - -class SqlServerLinkedService(LinkedService): - """SQL Server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param user_name: The on-premises Windows authentication user name. Type: string (or Expression - with resultType string). - :type user_name: any - :param password: The on-premises Windows authentication password. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlServerLinkedService, self).__init__(**kwargs) - self.type = 'SqlServer' # type: str - self.connection_string = kwargs['connection_string'] - self.user_name = kwargs.get('user_name', None) - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class SqlServerSink(CopySink): - """A copy activity SQL server sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param sql_writer_stored_procedure_name: SQL writer stored procedure name. Type: string (or - Expression with resultType string). - :type sql_writer_stored_procedure_name: any - :param sql_writer_table_type: SQL writer table type. Type: string (or Expression with - resultType string). - :type sql_writer_table_type: any - :param pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType - string). - :type pre_copy_script: any - :param stored_procedure_parameters: SQL stored procedure parameters. - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - :param stored_procedure_table_type_parameter_name: The stored procedure parameter name of the - table type. Type: string (or Expression with resultType string). - :type stored_procedure_table_type_parameter_name: any - :param table_option: The option to handle sink table, such as autoCreate. For now only - 'autoCreate' value is supported. Type: string (or Expression with resultType string). - :type table_option: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'sql_writer_stored_procedure_name': {'key': 'sqlWriterStoredProcedureName', 'type': 'object'}, - 'sql_writer_table_type': {'key': 'sqlWriterTableType', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - 'stored_procedure_table_type_parameter_name': {'key': 'storedProcedureTableTypeParameterName', 'type': 'object'}, - 'table_option': {'key': 'tableOption', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlServerSink, self).__init__(**kwargs) - self.type = 'SqlServerSink' # type: str - self.sql_writer_stored_procedure_name = kwargs.get('sql_writer_stored_procedure_name', None) - self.sql_writer_table_type = kwargs.get('sql_writer_table_type', None) - self.pre_copy_script = kwargs.get('pre_copy_script', None) - self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) - self.stored_procedure_table_type_parameter_name = kwargs.get('stored_procedure_table_type_parameter_name', None) - self.table_option = kwargs.get('table_option', None) - - -class SqlServerSource(TabularSource): - """A copy activity SQL server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). - :type sql_reader_query: any - :param sql_reader_stored_procedure_name: Name of the stored procedure for a SQL Database - source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression - with resultType string). - :type sql_reader_stored_procedure_name: any - :param stored_procedure_parameters: Value and type setting for stored procedure parameters. - Example: "{Parameter1: {value: "1", type: "int"}}". - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - :param produce_additional_types: Which additional types to produce. - :type produce_additional_types: any - :param partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". - :type partition_option: any - :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.SqlPartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, - 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - 'produce_additional_types': {'key': 'produceAdditionalTypes', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'object'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlServerSource, self).__init__(**kwargs) - self.type = 'SqlServerSource' # type: str - self.sql_reader_query = kwargs.get('sql_reader_query', None) - self.sql_reader_stored_procedure_name = kwargs.get('sql_reader_stored_procedure_name', None) - self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) - self.produce_additional_types = kwargs.get('produce_additional_types', None) - self.partition_option = kwargs.get('partition_option', None) - self.partition_settings = kwargs.get('partition_settings', None) - - -class SqlServerStoredProcedureActivity(ExecutionActivity): - """SQL stored procedure activity type. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param stored_procedure_name: Required. Stored procedure name. Type: string (or Expression with - resultType string). - :type stored_procedure_name: any - :param stored_procedure_parameters: Value and type setting for stored procedure parameters. - Example: "{Parameter1: {value: "1", type: "int"}}". - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'stored_procedure_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'stored_procedure_name': {'key': 'typeProperties.storedProcedureName', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'typeProperties.storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlServerStoredProcedureActivity, self).__init__(**kwargs) - self.type = 'SqlServerStoredProcedure' # type: str - self.stored_procedure_name = kwargs['stored_procedure_name'] - self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) - - -class SqlServerTableDataset(Dataset): - """The on-premises SQL Server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param schema_type_properties_schema: The schema name of the SQL Server dataset. Type: string - (or Expression with resultType string). - :type schema_type_properties_schema: any - :param table: The table name of the SQL Server dataset. Type: string (or Expression with - resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlServerTableDataset, self).__init__(**kwargs) - self.type = 'SqlServerTable' # type: str - self.table_name = kwargs.get('table_name', None) - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - self.table = kwargs.get('table', None) - - -class SqlSink(CopySink): - """A copy activity SQL sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param sql_writer_stored_procedure_name: SQL writer stored procedure name. Type: string (or - Expression with resultType string). - :type sql_writer_stored_procedure_name: any - :param sql_writer_table_type: SQL writer table type. Type: string (or Expression with - resultType string). - :type sql_writer_table_type: any - :param pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType - string). - :type pre_copy_script: any - :param stored_procedure_parameters: SQL stored procedure parameters. - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - :param stored_procedure_table_type_parameter_name: The stored procedure parameter name of the - table type. Type: string (or Expression with resultType string). - :type stored_procedure_table_type_parameter_name: any - :param table_option: The option to handle sink table, such as autoCreate. For now only - 'autoCreate' value is supported. Type: string (or Expression with resultType string). - :type table_option: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'sql_writer_stored_procedure_name': {'key': 'sqlWriterStoredProcedureName', 'type': 'object'}, - 'sql_writer_table_type': {'key': 'sqlWriterTableType', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - 'stored_procedure_table_type_parameter_name': {'key': 'storedProcedureTableTypeParameterName', 'type': 'object'}, - 'table_option': {'key': 'tableOption', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlSink, self).__init__(**kwargs) - self.type = 'SqlSink' # type: str - self.sql_writer_stored_procedure_name = kwargs.get('sql_writer_stored_procedure_name', None) - self.sql_writer_table_type = kwargs.get('sql_writer_table_type', None) - self.pre_copy_script = kwargs.get('pre_copy_script', None) - self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) - self.stored_procedure_table_type_parameter_name = kwargs.get('stored_procedure_table_type_parameter_name', None) - self.table_option = kwargs.get('table_option', None) - - -class SqlSource(TabularSource): - """A copy activity SQL source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). - :type sql_reader_query: any - :param sql_reader_stored_procedure_name: Name of the stored procedure for a SQL Database - source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression - with resultType string). - :type sql_reader_stored_procedure_name: any - :param stored_procedure_parameters: Value and type setting for stored procedure parameters. - Example: "{Parameter1: {value: "1", type: "int"}}". - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - :param isolation_level: Specifies the transaction locking behavior for the SQL source. Allowed - values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value - is ReadCommitted. Type: string (or Expression with resultType string). - :type isolation_level: any - :param partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". - :type partition_option: any - :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.SqlPartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, - 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - 'isolation_level': {'key': 'isolationLevel', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'object'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(SqlSource, self).__init__(**kwargs) - self.type = 'SqlSource' # type: str - self.sql_reader_query = kwargs.get('sql_reader_query', None) - self.sql_reader_stored_procedure_name = kwargs.get('sql_reader_stored_procedure_name', None) - self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) - self.isolation_level = kwargs.get('isolation_level', None) - self.partition_option = kwargs.get('partition_option', None) - self.partition_settings = kwargs.get('partition_settings', None) - - -class SquareLinkedService(LinkedService): - """Square Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_properties: Properties used to connect to Square. It is mutually exclusive - with any other properties in the linked service. Type: object. - :type connection_properties: any - :param host: Required. The URL of the Square instance. (i.e. mystore.mysquare.com). - :type host: any - :param client_id: Required. The client ID associated with your Square application. - :type client_id: any - :param client_secret: The client secret associated with your Square application. - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param redirect_uri: Required. The redirect URL assigned in the Square application dashboard. - (i.e. http://localhost:2500). - :type redirect_uri: any - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'client_id': {'required': True}, - 'redirect_uri': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_properties': {'key': 'typeProperties.connectionProperties', 'type': 'object'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'redirect_uri': {'key': 'typeProperties.redirectUri', 'type': 'object'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SquareLinkedService, self).__init__(**kwargs) - self.type = 'Square' # type: str - self.connection_properties = kwargs.get('connection_properties', None) - self.host = kwargs['host'] - self.client_id = kwargs['client_id'] - self.client_secret = kwargs.get('client_secret', None) - self.redirect_uri = kwargs['redirect_uri'] - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.use_host_verification = kwargs.get('use_host_verification', None) - self.use_peer_verification = kwargs.get('use_peer_verification', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class SquareObjectDataset(Dataset): - """Square Service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SquareObjectDataset, self).__init__(**kwargs) - self.type = 'SquareObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class SquareSource(TabularSource): - """A copy activity Square Service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SquareSource, self).__init__(**kwargs) - self.type = 'SquareSource' # type: str - self.query = kwargs.get('query', None) - - -class SSISAccessCredential(msrest.serialization.Model): - """SSIS access credential. - - All required parameters must be populated in order to send to Azure. - - :param domain: Required. Domain for windows authentication. - :type domain: any - :param user_name: Required. UseName for windows authentication. - :type user_name: any - :param password: Required. Password for windows authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - """ - - _validation = { - 'domain': {'required': True}, - 'user_name': {'required': True}, - 'password': {'required': True}, - } - - _attribute_map = { - 'domain': {'key': 'domain', 'type': 'object'}, - 'user_name': {'key': 'userName', 'type': 'object'}, - 'password': {'key': 'password', 'type': 'SecretBase'}, - } - - def __init__( - self, - **kwargs - ): - super(SSISAccessCredential, self).__init__(**kwargs) - self.domain = kwargs['domain'] - self.user_name = kwargs['user_name'] - self.password = kwargs['password'] - - -class SSISChildPackage(msrest.serialization.Model): - """SSIS embedded child package. - - All required parameters must be populated in order to send to Azure. - - :param package_path: Required. Path for embedded child package. Type: string (or Expression - with resultType string). - :type package_path: any - :param package_name: Name for embedded child package. - :type package_name: str - :param package_content: Required. Content for embedded child package. Type: string (or - Expression with resultType string). - :type package_content: any - :param package_last_modified_date: Last modified date for embedded child package. - :type package_last_modified_date: str - """ - - _validation = { - 'package_path': {'required': True}, - 'package_content': {'required': True}, - } - - _attribute_map = { - 'package_path': {'key': 'packagePath', 'type': 'object'}, - 'package_name': {'key': 'packageName', 'type': 'str'}, - 'package_content': {'key': 'packageContent', 'type': 'object'}, - 'package_last_modified_date': {'key': 'packageLastModifiedDate', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SSISChildPackage, self).__init__(**kwargs) - self.package_path = kwargs['package_path'] - self.package_name = kwargs.get('package_name', None) - self.package_content = kwargs['package_content'] - self.package_last_modified_date = kwargs.get('package_last_modified_date', None) - - -class SSISExecutionCredential(msrest.serialization.Model): - """SSIS package execution credential. - - All required parameters must be populated in order to send to Azure. - - :param domain: Required. Domain for windows authentication. - :type domain: any - :param user_name: Required. UseName for windows authentication. - :type user_name: any - :param password: Required. Password for windows authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecureString - """ - - _validation = { - 'domain': {'required': True}, - 'user_name': {'required': True}, - 'password': {'required': True}, - } - - _attribute_map = { - 'domain': {'key': 'domain', 'type': 'object'}, - 'user_name': {'key': 'userName', 'type': 'object'}, - 'password': {'key': 'password', 'type': 'SecureString'}, - } - - def __init__( - self, - **kwargs - ): - super(SSISExecutionCredential, self).__init__(**kwargs) - self.domain = kwargs['domain'] - self.user_name = kwargs['user_name'] - self.password = kwargs['password'] - - -class SSISExecutionParameter(msrest.serialization.Model): - """SSIS execution parameter. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. SSIS package execution parameter value. Type: string (or Expression - with resultType string). - :type value: any - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SSISExecutionParameter, self).__init__(**kwargs) - self.value = kwargs['value'] - - -class SSISLogLocation(msrest.serialization.Model): - """SSIS package execution log location. - - All required parameters must be populated in order to send to Azure. - - :param log_path: Required. The SSIS package execution log path. Type: string (or Expression - with resultType string). - :type log_path: any - :param type: Required. The type of SSIS log location. Possible values include: "File". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.SsisLogLocationType - :param access_credential: The package execution log access credential. - :type access_credential: ~azure.synapse.artifacts.v2020_12_01.models.SSISAccessCredential - :param log_refresh_interval: Specifies the interval to refresh log. The default interval is 5 - minutes. Type: string (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type log_refresh_interval: any - """ - - _validation = { - 'log_path': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'log_path': {'key': 'logPath', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'str'}, - 'access_credential': {'key': 'typeProperties.accessCredential', 'type': 'SSISAccessCredential'}, - 'log_refresh_interval': {'key': 'typeProperties.logRefreshInterval', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SSISLogLocation, self).__init__(**kwargs) - self.log_path = kwargs['log_path'] - self.type = kwargs['type'] - self.access_credential = kwargs.get('access_credential', None) - self.log_refresh_interval = kwargs.get('log_refresh_interval', None) - - -class SsisObjectMetadataStatusResponse(msrest.serialization.Model): - """The status of the operation. - - :param status: The status of the operation. - :type status: str - :param name: The operation name. - :type name: str - :param properties: The operation properties. - :type properties: str - :param error: The operation error message. - :type error: str - """ - - _attribute_map = { - 'status': {'key': 'status', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'str'}, - 'error': {'key': 'error', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SsisObjectMetadataStatusResponse, self).__init__(**kwargs) - self.status = kwargs.get('status', None) - self.name = kwargs.get('name', None) - self.properties = kwargs.get('properties', None) - self.error = kwargs.get('error', None) - - -class SSISPackageLocation(msrest.serialization.Model): - """SSIS package location. - - :param package_path: The SSIS package path. Type: string (or Expression with resultType - string). - :type package_path: any - :param type: The type of SSIS package location. Possible values include: "SSISDB", "File", - "InlinePackage", "PackageStore". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.SsisPackageLocationType - :param package_password: Password of the package. - :type package_password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param access_credential: The package access credential. - :type access_credential: ~azure.synapse.artifacts.v2020_12_01.models.SSISAccessCredential - :param configuration_path: The configuration file of the package execution. Type: string (or - Expression with resultType string). - :type configuration_path: any - :param configuration_access_credential: The configuration file access credential. - :type configuration_access_credential: - ~azure.synapse.artifacts.v2020_12_01.models.SSISAccessCredential - :param package_name: The package name. - :type package_name: str - :param package_content: The embedded package content. Type: string (or Expression with - resultType string). - :type package_content: any - :param package_last_modified_date: The embedded package last modified date. - :type package_last_modified_date: str - :param child_packages: The embedded child package list. - :type child_packages: list[~azure.synapse.artifacts.v2020_12_01.models.SSISChildPackage] - """ - - _attribute_map = { - 'package_path': {'key': 'packagePath', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'str'}, - 'package_password': {'key': 'typeProperties.packagePassword', 'type': 'SecretBase'}, - 'access_credential': {'key': 'typeProperties.accessCredential', 'type': 'SSISAccessCredential'}, - 'configuration_path': {'key': 'typeProperties.configurationPath', 'type': 'object'}, - 'configuration_access_credential': {'key': 'typeProperties.configurationAccessCredential', 'type': 'SSISAccessCredential'}, - 'package_name': {'key': 'typeProperties.packageName', 'type': 'str'}, - 'package_content': {'key': 'typeProperties.packageContent', 'type': 'object'}, - 'package_last_modified_date': {'key': 'typeProperties.packageLastModifiedDate', 'type': 'str'}, - 'child_packages': {'key': 'typeProperties.childPackages', 'type': '[SSISChildPackage]'}, - } - - def __init__( - self, - **kwargs - ): - super(SSISPackageLocation, self).__init__(**kwargs) - self.package_path = kwargs.get('package_path', None) - self.type = kwargs.get('type', None) - self.package_password = kwargs.get('package_password', None) - self.access_credential = kwargs.get('access_credential', None) - self.configuration_path = kwargs.get('configuration_path', None) - self.configuration_access_credential = kwargs.get('configuration_access_credential', None) - self.package_name = kwargs.get('package_name', None) - self.package_content = kwargs.get('package_content', None) - self.package_last_modified_date = kwargs.get('package_last_modified_date', None) - self.child_packages = kwargs.get('child_packages', None) - - -class SSISPropertyOverride(msrest.serialization.Model): - """SSIS property override. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. SSIS package property override value. Type: string (or Expression with - resultType string). - :type value: any - :param is_sensitive: Whether SSIS package property override value is sensitive data. Value will - be encrypted in SSISDB if it is true. - :type is_sensitive: bool - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': 'object'}, - 'is_sensitive': {'key': 'isSensitive', 'type': 'bool'}, - } - - def __init__( - self, - **kwargs - ): - super(SSISPropertyOverride, self).__init__(**kwargs) - self.value = kwargs['value'] - self.is_sensitive = kwargs.get('is_sensitive', None) - - -class StagingSettings(msrest.serialization.Model): - """Staging settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param linked_service_name: Required. Staging linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param path: The path to storage for storing the interim data. Type: string (or Expression with - resultType string). - :type path: any - :param enable_compression: Specifies whether to use compression when copying data via an - interim staging. Default value is false. Type: boolean (or Expression with resultType boolean). - :type enable_compression: any - """ - - _validation = { - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'path': {'key': 'path', 'type': 'object'}, - 'enable_compression': {'key': 'enableCompression', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(StagingSettings, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.linked_service_name = kwargs['linked_service_name'] - self.path = kwargs.get('path', None) - self.enable_compression = kwargs.get('enable_compression', None) - - -class StartDataFlowDebugSessionRequest(msrest.serialization.Model): - """Request body structure for starting data flow debug session. - - :param session_id: The ID of data flow debug session. - :type session_id: str - :param data_flow: Data flow instance. - :type data_flow: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowResource - :param datasets: List of datasets. - :type datasets: list[~azure.synapse.artifacts.v2020_12_01.models.DatasetResource] - :param linked_services: List of linked services. - :type linked_services: list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceResource] - :param staging: Staging info for debug session. - :type staging: any - :param debug_settings: Data flow debug settings. - :type debug_settings: any - :param incremental_debug: The type of new Databricks cluster. - :type incremental_debug: bool - """ - - _attribute_map = { - 'session_id': {'key': 'sessionId', 'type': 'str'}, - 'data_flow': {'key': 'dataFlow', 'type': 'DataFlowResource'}, - 'datasets': {'key': 'datasets', 'type': '[DatasetResource]'}, - 'linked_services': {'key': 'linkedServices', 'type': '[LinkedServiceResource]'}, - 'staging': {'key': 'staging', 'type': 'object'}, - 'debug_settings': {'key': 'debugSettings', 'type': 'object'}, - 'incremental_debug': {'key': 'incrementalDebug', 'type': 'bool'}, - } - - def __init__( - self, - **kwargs - ): - super(StartDataFlowDebugSessionRequest, self).__init__(**kwargs) - self.session_id = kwargs.get('session_id', None) - self.data_flow = kwargs.get('data_flow', None) - self.datasets = kwargs.get('datasets', None) - self.linked_services = kwargs.get('linked_services', None) - self.staging = kwargs.get('staging', None) - self.debug_settings = kwargs.get('debug_settings', None) - self.incremental_debug = kwargs.get('incremental_debug', None) - - -class StartDataFlowDebugSessionResponse(msrest.serialization.Model): - """Response body structure for starting data flow debug session. - - :param job_version: The ID of data flow debug job version. - :type job_version: str - """ - - _attribute_map = { - 'job_version': {'key': 'jobVersion', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(StartDataFlowDebugSessionResponse, self).__init__(**kwargs) - self.job_version = kwargs.get('job_version', None) - - -class StoredProcedureParameter(msrest.serialization.Model): - """SQL stored procedure parameter. - - :param value: Stored procedure parameter value. Type: string (or Expression with resultType - string). - :type value: any - :param type: Stored procedure parameter type. Possible values include: "String", "Int", - "Int64", "Decimal", "Guid", "Boolean", "Date". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameterType - """ - - _attribute_map = { - 'value': {'key': 'value', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(StoredProcedureParameter, self).__init__(**kwargs) - self.value = kwargs.get('value', None) - self.type = kwargs.get('type', None) - - -class SwitchActivity(ControlActivity): - """This activity evaluates an expression and executes activities under the cases property that correspond to the expression evaluation expected in the equals property. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param on: Required. An expression that would evaluate to a string or integer. This is used to - determine the block of activities in cases that will be executed. - :type on: ~azure.synapse.artifacts.v2020_12_01.models.Expression - :param cases: List of cases that correspond to expected values of the 'on' property. This is an - optional property and if not provided, the activity will execute activities provided in - defaultActivities. - :type cases: list[~azure.synapse.artifacts.v2020_12_01.models.SwitchCase] - :param default_activities: List of activities to execute if no case condition is satisfied. - This is an optional property and if not provided, the activity will exit without any action. - :type default_activities: list[~azure.synapse.artifacts.v2020_12_01.models.Activity] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'on': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'on': {'key': 'typeProperties.on', 'type': 'Expression'}, - 'cases': {'key': 'typeProperties.cases', 'type': '[SwitchCase]'}, - 'default_activities': {'key': 'typeProperties.defaultActivities', 'type': '[Activity]'}, - } - - def __init__( - self, - **kwargs - ): - super(SwitchActivity, self).__init__(**kwargs) - self.type = 'Switch' # type: str - self.on = kwargs['on'] - self.cases = kwargs.get('cases', None) - self.default_activities = kwargs.get('default_activities', None) - - -class SwitchCase(msrest.serialization.Model): - """Switch cases with have a value and corresponding activities. - - :param value: Expected value that satisfies the expression result of the 'on' property. - :type value: str - :param activities: List of activities to execute for satisfied case condition. - :type activities: list[~azure.synapse.artifacts.v2020_12_01.models.Activity] - """ - - _attribute_map = { - 'value': {'key': 'value', 'type': 'str'}, - 'activities': {'key': 'activities', 'type': '[Activity]'}, - } - - def __init__( - self, - **kwargs - ): - super(SwitchCase, self).__init__(**kwargs) - self.value = kwargs.get('value', None) - self.activities = kwargs.get('activities', None) - - -class SybaseLinkedService(LinkedService): - """Linked service for Sybase data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param server: Required. Server name for connection. Type: string (or Expression with - resultType string). - :type server: any - :param database: Required. Database name for connection. Type: string (or Expression with - resultType string). - :type database: any - :param schema: Schema name for connection. Type: string (or Expression with resultType string). - :type schema: any - :param authentication_type: AuthenticationType to be used for connection. Possible values - include: "Basic", "Windows". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.SybaseAuthenticationType - :param username: Username for authentication. Type: string (or Expression with resultType - string). - :type username: any - :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'server': {'required': True}, - 'database': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SybaseLinkedService, self).__init__(**kwargs) - self.type = 'Sybase' # type: str - self.server = kwargs['server'] - self.database = kwargs['database'] - self.schema = kwargs.get('schema', None) - self.authentication_type = kwargs.get('authentication_type', None) - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class SybaseSource(TabularSource): - """A copy activity source for Sybase databases. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SybaseSource, self).__init__(**kwargs) - self.type = 'SybaseSource' # type: str - self.query = kwargs.get('query', None) - - -class SybaseTableDataset(Dataset): - """The Sybase table dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The Sybase table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(SybaseTableDataset, self).__init__(**kwargs) - self.type = 'SybaseTable' # type: str - self.table_name = kwargs.get('table_name', None) - - -class SynapseNotebookActivity(ExecutionActivity): - """Execute Synapse notebook activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param notebook: Required. Synapse notebook reference. - :type notebook: ~azure.synapse.artifacts.v2020_12_01.models.SynapseNotebookReference - :param parameters: Notebook parameters. - :type parameters: dict[str, any] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'notebook': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'notebook': {'key': 'typeProperties.notebook', 'type': 'SynapseNotebookReference'}, - 'parameters': {'key': 'typeProperties.parameters', 'type': '{object}'}, - } - - def __init__( - self, - **kwargs - ): - super(SynapseNotebookActivity, self).__init__(**kwargs) - self.type = 'SynapseNotebook' # type: str - self.notebook = kwargs['notebook'] - self.parameters = kwargs.get('parameters', None) - - -class SynapseNotebookReference(msrest.serialization.Model): - """Synapse notebook reference type. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Synapse notebook reference type. Possible values include: - "NotebookReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.NotebookReferenceType - :param reference_name: Required. Reference notebook name. - :type reference_name: str - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SynapseNotebookReference, self).__init__(**kwargs) - self.type = kwargs['type'] - self.reference_name = kwargs['reference_name'] - - -class SynapseSparkJobDefinitionActivity(ExecutionActivity): - """Execute spark job activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param spark_job: Required. Synapse spark job reference. - :type spark_job: ~azure.synapse.artifacts.v2020_12_01.models.SynapseSparkJobReference - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'spark_job': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'spark_job': {'key': 'typeProperties.sparkJob', 'type': 'SynapseSparkJobReference'}, - } - - def __init__( - self, - **kwargs - ): - super(SynapseSparkJobDefinitionActivity, self).__init__(**kwargs) - self.type = 'SparkJob' # type: str - self.spark_job = kwargs['spark_job'] - - -class SynapseSparkJobReference(msrest.serialization.Model): - """Synapse spark job reference type. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Synapse spark job reference type. Possible values include: - "SparkJobDefinitionReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.SparkJobReferenceType - :param reference_name: Required. Reference spark job name. - :type reference_name: str - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SynapseSparkJobReference, self).__init__(**kwargs) - self.type = kwargs['type'] - self.reference_name = kwargs['reference_name'] - - -class TabularTranslator(CopyTranslator): - """A copy activity tabular translator. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy translator type.Constant filled by server. - :type type: str - :param column_mappings: Column mappings. Example: "UserId: MyUserId, Group: MyGroup, Name: - MyName" Type: string (or Expression with resultType string). This property will be retired. - Please use mappings property. - :type column_mappings: any - :param schema_mapping: The schema mapping to map between tabular data and hierarchical data. - Example: {"Column1": "$.Column1", "Column2": "$.Column2.Property1", "Column3": - "$.Column2.Property2"}. Type: object (or Expression with resultType object). This property will - be retired. Please use mappings property. - :type schema_mapping: any - :param collection_reference: The JSON Path of the Nested Array that is going to do cross-apply. - Type: object (or Expression with resultType object). - :type collection_reference: any - :param map_complex_values_to_string: Whether to map complex (array and object) values to simple - strings in json format. Type: boolean (or Expression with resultType boolean). - :type map_complex_values_to_string: any - :param mappings: Column mappings with logical types. Tabular->tabular example: - [{"source":{"name":"CustomerName","type":"String"},"sink":{"name":"ClientName","type":"String"}},{"source":{"name":"CustomerAddress","type":"String"},"sink":{"name":"ClientAddress","type":"String"}}]. - Hierarchical->tabular example: - [{"source":{"path":"$.CustomerName","type":"String"},"sink":{"name":"ClientName","type":"String"}},{"source":{"path":"$.CustomerAddress","type":"String"},"sink":{"name":"ClientAddress","type":"String"}}]. - Type: object (or Expression with resultType object). - :type mappings: any - :param type_conversion: Whether to enable the advanced type conversion feature in the Copy - activity. Type: boolean (or Expression with resultType boolean). - :type type_conversion: any - :param type_conversion_settings: Type conversion settings. - :type type_conversion_settings: - ~azure.synapse.artifacts.v2020_12_01.models.TypeConversionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'column_mappings': {'key': 'columnMappings', 'type': 'object'}, - 'schema_mapping': {'key': 'schemaMapping', 'type': 'object'}, - 'collection_reference': {'key': 'collectionReference', 'type': 'object'}, - 'map_complex_values_to_string': {'key': 'mapComplexValuesToString', 'type': 'object'}, - 'mappings': {'key': 'mappings', 'type': 'object'}, - 'type_conversion': {'key': 'typeConversion', 'type': 'object'}, - 'type_conversion_settings': {'key': 'typeConversionSettings', 'type': 'TypeConversionSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(TabularTranslator, self).__init__(**kwargs) - self.type = 'TabularTranslator' # type: str - self.column_mappings = kwargs.get('column_mappings', None) - self.schema_mapping = kwargs.get('schema_mapping', None) - self.collection_reference = kwargs.get('collection_reference', None) - self.map_complex_values_to_string = kwargs.get('map_complex_values_to_string', None) - self.mappings = kwargs.get('mappings', None) - self.type_conversion = kwargs.get('type_conversion', None) - self.type_conversion_settings = kwargs.get('type_conversion_settings', None) - - -class TarGZipReadSettings(CompressionReadSettings): - """The TarGZip compression read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The Compression setting type.Constant filled by server. - :type type: str - :param preserve_compression_file_name_as_folder: Preserve the compression file name as folder - path. Type: boolean (or Expression with resultType boolean). - :type preserve_compression_file_name_as_folder: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'preserve_compression_file_name_as_folder': {'key': 'preserveCompressionFileNameAsFolder', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(TarGZipReadSettings, self).__init__(**kwargs) - self.type = 'TarGZipReadSettings' # type: str - self.preserve_compression_file_name_as_folder = kwargs.get('preserve_compression_file_name_as_folder', None) - - -class TarReadSettings(CompressionReadSettings): - """The Tar compression read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The Compression setting type.Constant filled by server. - :type type: str - :param preserve_compression_file_name_as_folder: Preserve the compression file name as folder - path. Type: boolean (or Expression with resultType boolean). - :type preserve_compression_file_name_as_folder: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'preserve_compression_file_name_as_folder': {'key': 'preserveCompressionFileNameAsFolder', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(TarReadSettings, self).__init__(**kwargs) - self.type = 'TarReadSettings' # type: str - self.preserve_compression_file_name_as_folder = kwargs.get('preserve_compression_file_name_as_folder', None) - - -class TeradataLinkedService(LinkedService): - """Linked service for Teradata data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Teradata ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param server: Server name for connection. Type: string (or Expression with resultType string). - :type server: any - :param authentication_type: AuthenticationType to be used for connection. Possible values - include: "Basic", "Windows". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.TeradataAuthenticationType - :param username: Username for authentication. Type: string (or Expression with resultType - string). - :type username: any - :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(TeradataLinkedService, self).__init__(**kwargs) - self.type = 'Teradata' # type: str - self.connection_string = kwargs.get('connection_string', None) - self.server = kwargs.get('server', None) - self.authentication_type = kwargs.get('authentication_type', None) - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class TeradataPartitionSettings(msrest.serialization.Model): - """The settings that will be leveraged for teradata source partitioning. - - :param partition_column_name: The name of the column that will be used for proceeding range or - hash partitioning. Type: string (or Expression with resultType string). - :type partition_column_name: any - :param partition_upper_bound: The maximum value of column specified in partitionColumnName that - will be used for proceeding range partitioning. Type: string (or Expression with resultType - string). - :type partition_upper_bound: any - :param partition_lower_bound: The minimum value of column specified in partitionColumnName that - will be used for proceeding range partitioning. Type: string (or Expression with resultType - string). - :type partition_lower_bound: any - """ - - _attribute_map = { - 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, - 'partition_upper_bound': {'key': 'partitionUpperBound', 'type': 'object'}, - 'partition_lower_bound': {'key': 'partitionLowerBound', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(TeradataPartitionSettings, self).__init__(**kwargs) - self.partition_column_name = kwargs.get('partition_column_name', None) - self.partition_upper_bound = kwargs.get('partition_upper_bound', None) - self.partition_lower_bound = kwargs.get('partition_lower_bound', None) - - -class TeradataSource(TabularSource): - """A copy activity Teradata source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Teradata query. Type: string (or Expression with resultType string). - :type query: any - :param partition_option: The partition mechanism that will be used for teradata read in - parallel. Possible values include: "None", "Hash", "DynamicRange". - :type partition_option: str or - ~azure.synapse.artifacts.v2020_12_01.models.TeradataPartitionOption - :param partition_settings: The settings that will be leveraged for teradata source - partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.TeradataPartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'TeradataPartitionSettings'}, - } - - def __init__( - self, - **kwargs - ): - super(TeradataSource, self).__init__(**kwargs) - self.type = 'TeradataSource' # type: str - self.query = kwargs.get('query', None) - self.partition_option = kwargs.get('partition_option', None) - self.partition_settings = kwargs.get('partition_settings', None) - - -class TeradataTableDataset(Dataset): - """The Teradata database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param database: The database name of Teradata. Type: string (or Expression with resultType - string). - :type database: any - :param table: The table name of Teradata. Type: string (or Expression with resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(TeradataTableDataset, self).__init__(**kwargs) - self.type = 'TeradataTable' # type: str - self.database = kwargs.get('database', None) - self.table = kwargs.get('table', None) - - -class TextFormat(DatasetStorageFormat): - """The data stored in text format. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage format.Constant filled by server. - :type type: str - :param serializer: Serializer. Type: string (or Expression with resultType string). - :type serializer: any - :param deserializer: Deserializer. Type: string (or Expression with resultType string). - :type deserializer: any - :param column_delimiter: The column delimiter. Type: string (or Expression with resultType - string). - :type column_delimiter: any - :param row_delimiter: The row delimiter. Type: string (or Expression with resultType string). - :type row_delimiter: any - :param escape_char: The escape character. Type: string (or Expression with resultType string). - :type escape_char: any - :param quote_char: The quote character. Type: string (or Expression with resultType string). - :type quote_char: any - :param null_value: The null value string. Type: string (or Expression with resultType string). - :type null_value: any - :param encoding_name: The code page name of the preferred encoding. If miss, the default value - is ΓÇ£utf-8ΓÇ¥, unless BOM denotes another Unicode encoding. Refer to the ΓÇ£NameΓÇ¥ column of - the table in the following link to set supported values: - https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string (or Expression with - resultType string). - :type encoding_name: any - :param treat_empty_as_null: Treat empty column values in the text file as null. The default - value is true. Type: boolean (or Expression with resultType boolean). - :type treat_empty_as_null: any - :param skip_line_count: The number of lines/rows to be skipped when parsing text files. The - default value is 0. Type: integer (or Expression with resultType integer). - :type skip_line_count: any - :param first_row_as_header: When used as input, treat the first row of data as headers. When - used as output,write the headers into the output as the first row of data. The default value is - false. Type: boolean (or Expression with resultType boolean). - :type first_row_as_header: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'serializer': {'key': 'serializer', 'type': 'object'}, - 'deserializer': {'key': 'deserializer', 'type': 'object'}, - 'column_delimiter': {'key': 'columnDelimiter', 'type': 'object'}, - 'row_delimiter': {'key': 'rowDelimiter', 'type': 'object'}, - 'escape_char': {'key': 'escapeChar', 'type': 'object'}, - 'quote_char': {'key': 'quoteChar', 'type': 'object'}, - 'null_value': {'key': 'nullValue', 'type': 'object'}, - 'encoding_name': {'key': 'encodingName', 'type': 'object'}, - 'treat_empty_as_null': {'key': 'treatEmptyAsNull', 'type': 'object'}, - 'skip_line_count': {'key': 'skipLineCount', 'type': 'object'}, - 'first_row_as_header': {'key': 'firstRowAsHeader', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(TextFormat, self).__init__(**kwargs) - self.type = 'TextFormat' # type: str - self.column_delimiter = kwargs.get('column_delimiter', None) - self.row_delimiter = kwargs.get('row_delimiter', None) - self.escape_char = kwargs.get('escape_char', None) - self.quote_char = kwargs.get('quote_char', None) - self.null_value = kwargs.get('null_value', None) - self.encoding_name = kwargs.get('encoding_name', None) - self.treat_empty_as_null = kwargs.get('treat_empty_as_null', None) - self.skip_line_count = kwargs.get('skip_line_count', None) - self.first_row_as_header = kwargs.get('first_row_as_header', None) - - -class TriggerDependencyProvisioningStatus(msrest.serialization.Model): - """Defines the response of a provision trigger dependency operation. - - All required parameters must be populated in order to send to Azure. - - :param trigger_name: Required. Trigger name. - :type trigger_name: str - :param provisioning_status: Required. Provisioning status. - :type provisioning_status: str - """ - - _validation = { - 'trigger_name': {'required': True}, - 'provisioning_status': {'required': True}, - } - - _attribute_map = { - 'trigger_name': {'key': 'triggerName', 'type': 'str'}, - 'provisioning_status': {'key': 'provisioningStatus', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(TriggerDependencyProvisioningStatus, self).__init__(**kwargs) - self.trigger_name = kwargs['trigger_name'] - self.provisioning_status = kwargs['provisioning_status'] - - -class TriggerDependencyReference(DependencyReference): - """Trigger referenced dependency. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: TumblingWindowTriggerDependencyReference. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. The type of dependency reference.Constant filled by server. - :type type: str - :param reference_trigger: Required. Referenced trigger. - :type reference_trigger: ~azure.synapse.artifacts.v2020_12_01.models.TriggerReference - """ - - _validation = { - 'type': {'required': True}, - 'reference_trigger': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_trigger': {'key': 'referenceTrigger', 'type': 'TriggerReference'}, - } - - _subtype_map = { - 'type': {'TumblingWindowTriggerDependencyReference': 'TumblingWindowTriggerDependencyReference'} - } - - def __init__( - self, - **kwargs - ): - super(TriggerDependencyReference, self).__init__(**kwargs) - self.type = 'TriggerDependencyReference' # type: str - self.reference_trigger = kwargs['reference_trigger'] - - -class TriggerListResponse(msrest.serialization.Model): - """A list of trigger resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of triggers. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.TriggerResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[TriggerResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(TriggerListResponse, self).__init__(**kwargs) - self.value = kwargs['value'] - self.next_link = kwargs.get('next_link', None) - - -class TriggerPipelineReference(msrest.serialization.Model): - """Pipeline that needs to be triggered with the given parameters. - - :param pipeline_reference: Pipeline reference. - :type pipeline_reference: ~azure.synapse.artifacts.v2020_12_01.models.PipelineReference - :param parameters: Pipeline parameters. - :type parameters: dict[str, any] - """ - - _attribute_map = { - 'pipeline_reference': {'key': 'pipelineReference', 'type': 'PipelineReference'}, - 'parameters': {'key': 'parameters', 'type': '{object}'}, - } - - def __init__( - self, - **kwargs - ): - super(TriggerPipelineReference, self).__init__(**kwargs) - self.pipeline_reference = kwargs.get('pipeline_reference', None) - self.parameters = kwargs.get('parameters', None) - - -class TriggerReference(msrest.serialization.Model): - """Trigger reference type. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Trigger reference type. Possible values include: "TriggerReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerReferenceType - :param reference_name: Required. Reference trigger name. - :type reference_name: str - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(TriggerReference, self).__init__(**kwargs) - self.type = kwargs['type'] - self.reference_name = kwargs['reference_name'] - - -class TriggerResource(SubResource): - """Trigger resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Properties of the trigger. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.Trigger - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'Trigger'}, - } - - def __init__( - self, - **kwargs - ): - super(TriggerResource, self).__init__(**kwargs) - self.properties = kwargs['properties'] - - -class TriggerRun(msrest.serialization.Model): - """Trigger runs. - - Variables are only populated by the server, and will be ignored when sending a request. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :ivar trigger_run_id: Trigger run id. - :vartype trigger_run_id: str - :ivar trigger_name: Trigger name. - :vartype trigger_name: str - :ivar trigger_type: Trigger type. - :vartype trigger_type: str - :ivar trigger_run_timestamp: Trigger run start time. - :vartype trigger_run_timestamp: ~datetime.datetime - :ivar status: Trigger run status. Possible values include: "Succeeded", "Failed", "Inprogress". - :vartype status: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRunStatus - :ivar message: Trigger error message. - :vartype message: str - :ivar properties: List of property name and value related to trigger run. Name, value pair - depends on type of trigger. - :vartype properties: dict[str, str] - :ivar triggered_pipelines: List of pipeline name and run Id triggered by the trigger run. - :vartype triggered_pipelines: dict[str, str] - """ - - _validation = { - 'trigger_run_id': {'readonly': True}, - 'trigger_name': {'readonly': True}, - 'trigger_type': {'readonly': True}, - 'trigger_run_timestamp': {'readonly': True}, - 'status': {'readonly': True}, - 'message': {'readonly': True}, - 'properties': {'readonly': True}, - 'triggered_pipelines': {'readonly': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'trigger_run_id': {'key': 'triggerRunId', 'type': 'str'}, - 'trigger_name': {'key': 'triggerName', 'type': 'str'}, - 'trigger_type': {'key': 'triggerType', 'type': 'str'}, - 'trigger_run_timestamp': {'key': 'triggerRunTimestamp', 'type': 'iso-8601'}, - 'status': {'key': 'status', 'type': 'str'}, - 'message': {'key': 'message', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': '{str}'}, - 'triggered_pipelines': {'key': 'triggeredPipelines', 'type': '{str}'}, - } - - def __init__( - self, - **kwargs - ): - super(TriggerRun, self).__init__(**kwargs) - self.additional_properties = kwargs.get('additional_properties', None) - self.trigger_run_id = None - self.trigger_name = None - self.trigger_type = None - self.trigger_run_timestamp = None - self.status = None - self.message = None - self.properties = None - self.triggered_pipelines = None - - -class TriggerRunsQueryResponse(msrest.serialization.Model): - """A list of trigger runs. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of trigger runs. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.TriggerRun] - :param continuation_token: The continuation token for getting the next page of results, if any - remaining results exist, null otherwise. - :type continuation_token: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[TriggerRun]'}, - 'continuation_token': {'key': 'continuationToken', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(TriggerRunsQueryResponse, self).__init__(**kwargs) - self.value = kwargs['value'] - self.continuation_token = kwargs.get('continuation_token', None) - - -class TriggerSubscriptionOperationStatus(msrest.serialization.Model): - """Defines the response of a trigger subscription operation. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar trigger_name: Trigger name. - :vartype trigger_name: str - :ivar status: Event Subscription Status. Possible values include: "Enabled", "Provisioning", - "Deprovisioning", "Disabled", "Unknown". - :vartype status: str or ~azure.synapse.artifacts.v2020_12_01.models.EventSubscriptionStatus - """ - - _validation = { - 'trigger_name': {'readonly': True}, - 'status': {'readonly': True}, - } - - _attribute_map = { - 'trigger_name': {'key': 'triggerName', 'type': 'str'}, - 'status': {'key': 'status', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(TriggerSubscriptionOperationStatus, self).__init__(**kwargs) - self.trigger_name = None - self.status = None - - -class TumblingWindowTrigger(Trigger): - """Trigger that schedules pipeline runs for all fixed time interval windows from a start time without gaps and also supports backfill scenarios (when start time is in the past). - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Trigger type.Constant filled by server. - :type type: str - :param description: Trigger description. - :type description: str - :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are - called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRuntimeState - :param annotations: List of tags that can be used for describing the trigger. - :type annotations: list[any] - :param pipeline: Required. Pipeline for which runs are created when an event is fired for - trigger window that is ready. - :type pipeline: ~azure.synapse.artifacts.v2020_12_01.models.TriggerPipelineReference - :param frequency: Required. The frequency of the time windows. Possible values include: - "Minute", "Hour", "Month". - :type frequency: str or ~azure.synapse.artifacts.v2020_12_01.models.TumblingWindowFrequency - :param interval: Required. The interval of the time windows. The minimum interval allowed is 15 - Minutes. - :type interval: int - :param start_time: Required. The start time for the time period for the trigger during which - events are fired for windows that are ready. Only UTC time is currently supported. - :type start_time: ~datetime.datetime - :param end_time: The end time for the time period for the trigger during which events are fired - for windows that are ready. Only UTC time is currently supported. - :type end_time: ~datetime.datetime - :param delay: Specifies how long the trigger waits past due time before triggering new run. It - doesn't alter window start and end time. The default is 0. Type: string (or Expression with - resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type delay: any - :param max_concurrency: Required. The max number of parallel time windows (ready for execution) - for which a new run is triggered. - :type max_concurrency: int - :param retry_policy: Retry policy that will be applied for failed pipeline runs. - :type retry_policy: ~azure.synapse.artifacts.v2020_12_01.models.RetryPolicy - :param depends_on: Triggers that this trigger depends on. Only tumbling window triggers are - supported. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.DependencyReference] - """ - - _validation = { - 'type': {'required': True}, - 'runtime_state': {'readonly': True}, - 'pipeline': {'required': True}, - 'frequency': {'required': True}, - 'interval': {'required': True}, - 'start_time': {'required': True}, - 'max_concurrency': {'required': True, 'maximum': 50, 'minimum': 1}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'pipeline': {'key': 'pipeline', 'type': 'TriggerPipelineReference'}, - 'frequency': {'key': 'typeProperties.frequency', 'type': 'str'}, - 'interval': {'key': 'typeProperties.interval', 'type': 'int'}, - 'start_time': {'key': 'typeProperties.startTime', 'type': 'iso-8601'}, - 'end_time': {'key': 'typeProperties.endTime', 'type': 'iso-8601'}, - 'delay': {'key': 'typeProperties.delay', 'type': 'object'}, - 'max_concurrency': {'key': 'typeProperties.maxConcurrency', 'type': 'int'}, - 'retry_policy': {'key': 'typeProperties.retryPolicy', 'type': 'RetryPolicy'}, - 'depends_on': {'key': 'typeProperties.dependsOn', 'type': '[DependencyReference]'}, - } - - def __init__( - self, - **kwargs - ): - super(TumblingWindowTrigger, self).__init__(**kwargs) - self.type = 'TumblingWindowTrigger' # type: str - self.pipeline = kwargs['pipeline'] - self.frequency = kwargs['frequency'] - self.interval = kwargs['interval'] - self.start_time = kwargs['start_time'] - self.end_time = kwargs.get('end_time', None) - self.delay = kwargs.get('delay', None) - self.max_concurrency = kwargs['max_concurrency'] - self.retry_policy = kwargs.get('retry_policy', None) - self.depends_on = kwargs.get('depends_on', None) - - -class TumblingWindowTriggerDependencyReference(TriggerDependencyReference): - """Referenced tumbling window trigger dependency. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. The type of dependency reference.Constant filled by server. - :type type: str - :param reference_trigger: Required. Referenced trigger. - :type reference_trigger: ~azure.synapse.artifacts.v2020_12_01.models.TriggerReference - :param offset: Timespan applied to the start time of a tumbling window when evaluating - dependency. - :type offset: str - :param size: The size of the window when evaluating the dependency. If undefined the frequency - of the tumbling window will be used. - :type size: str - """ - - _validation = { - 'type': {'required': True}, - 'reference_trigger': {'required': True}, - 'offset': {'max_length': 15, 'min_length': 8, 'pattern': r'-?((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))'}, - 'size': {'max_length': 15, 'min_length': 8, 'pattern': r'((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))'}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_trigger': {'key': 'referenceTrigger', 'type': 'TriggerReference'}, - 'offset': {'key': 'offset', 'type': 'str'}, - 'size': {'key': 'size', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(TumblingWindowTriggerDependencyReference, self).__init__(**kwargs) - self.type = 'TumblingWindowTriggerDependencyReference' # type: str - self.offset = kwargs.get('offset', None) - self.size = kwargs.get('size', None) - - -class TypeConversionSettings(msrest.serialization.Model): - """Type conversion settings. - - :param allow_data_truncation: Whether to allow data truncation when converting the data. Type: - boolean (or Expression with resultType boolean). - :type allow_data_truncation: any - :param treat_boolean_as_number: Whether to treat boolean values as numbers. Type: boolean (or - Expression with resultType boolean). - :type treat_boolean_as_number: any - :param date_time_format: The format for DateTime values. Type: string (or Expression with - resultType string). - :type date_time_format: any - :param date_time_offset_format: The format for DateTimeOffset values. Type: string (or - Expression with resultType string). - :type date_time_offset_format: any - :param time_span_format: The format for TimeSpan values. Type: string (or Expression with - resultType string). - :type time_span_format: any - :param culture: The culture used to convert data from/to string. Type: string (or Expression - with resultType string). - :type culture: any - """ - - _attribute_map = { - 'allow_data_truncation': {'key': 'allowDataTruncation', 'type': 'object'}, - 'treat_boolean_as_number': {'key': 'treatBooleanAsNumber', 'type': 'object'}, - 'date_time_format': {'key': 'dateTimeFormat', 'type': 'object'}, - 'date_time_offset_format': {'key': 'dateTimeOffsetFormat', 'type': 'object'}, - 'time_span_format': {'key': 'timeSpanFormat', 'type': 'object'}, - 'culture': {'key': 'culture', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(TypeConversionSettings, self).__init__(**kwargs) - self.allow_data_truncation = kwargs.get('allow_data_truncation', None) - self.treat_boolean_as_number = kwargs.get('treat_boolean_as_number', None) - self.date_time_format = kwargs.get('date_time_format', None) - self.date_time_offset_format = kwargs.get('date_time_offset_format', None) - self.time_span_format = kwargs.get('time_span_format', None) - self.culture = kwargs.get('culture', None) - - -class UntilActivity(ControlActivity): - """This activity executes inner activities until the specified boolean expression results to true or timeout is reached, whichever is earlier. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param expression: Required. An expression that would evaluate to Boolean. The loop will - continue until this expression evaluates to true. - :type expression: ~azure.synapse.artifacts.v2020_12_01.models.Expression - :param timeout: Specifies the timeout for the activity to run. If there is no value specified, - it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or - Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). Type: string (or Expression with - resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type timeout: any - :param activities: Required. List of activities to execute. - :type activities: list[~azure.synapse.artifacts.v2020_12_01.models.Activity] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'expression': {'required': True}, - 'activities': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'expression': {'key': 'typeProperties.expression', 'type': 'Expression'}, - 'timeout': {'key': 'typeProperties.timeout', 'type': 'object'}, - 'activities': {'key': 'typeProperties.activities', 'type': '[Activity]'}, - } - - def __init__( - self, - **kwargs - ): - super(UntilActivity, self).__init__(**kwargs) - self.type = 'Until' # type: str - self.expression = kwargs['expression'] - self.timeout = kwargs.get('timeout', None) - self.activities = kwargs['activities'] - - -class UserProperty(msrest.serialization.Model): - """User property. - - All required parameters must be populated in order to send to Azure. - - :param name: Required. User property name. - :type name: str - :param value: Required. User property value. Type: string (or Expression with resultType - string). - :type value: any - """ - - _validation = { - 'name': {'required': True}, - 'value': {'required': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'value': {'key': 'value', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(UserProperty, self).__init__(**kwargs) - self.name = kwargs['name'] - self.value = kwargs['value'] - - -class ValidationActivity(ControlActivity): - """This activity verifies that an external resource exists. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param timeout: Specifies the timeout for the activity to run. If there is no value specified, - it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or - Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type timeout: any - :param sleep: A delay in seconds between validation attempts. If no value is specified, 10 - seconds will be used as the default. Type: integer (or Expression with resultType integer). - :type sleep: any - :param minimum_size: Can be used if dataset points to a file. The file must be greater than or - equal in size to the value specified. Type: integer (or Expression with resultType integer). - :type minimum_size: any - :param child_items: Can be used if dataset points to a folder. If set to true, the folder must - have at least one file. If set to false, the folder must be empty. Type: boolean (or Expression - with resultType boolean). - :type child_items: any - :param dataset: Required. Validation activity dataset reference. - :type dataset: ~azure.synapse.artifacts.v2020_12_01.models.DatasetReference - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'dataset': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'timeout': {'key': 'typeProperties.timeout', 'type': 'object'}, - 'sleep': {'key': 'typeProperties.sleep', 'type': 'object'}, - 'minimum_size': {'key': 'typeProperties.minimumSize', 'type': 'object'}, - 'child_items': {'key': 'typeProperties.childItems', 'type': 'object'}, - 'dataset': {'key': 'typeProperties.dataset', 'type': 'DatasetReference'}, - } - - def __init__( - self, - **kwargs - ): - super(ValidationActivity, self).__init__(**kwargs) - self.type = 'Validation' # type: str - self.timeout = kwargs.get('timeout', None) - self.sleep = kwargs.get('sleep', None) - self.minimum_size = kwargs.get('minimum_size', None) - self.child_items = kwargs.get('child_items', None) - self.dataset = kwargs['dataset'] - - -class VariableSpecification(msrest.serialization.Model): - """Definition of a single variable for a Pipeline. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Variable type. Possible values include: "String", "Bool", "Boolean", - "Array". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.VariableType - :param default_value: Default value of variable. - :type default_value: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'default_value': {'key': 'defaultValue', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(VariableSpecification, self).__init__(**kwargs) - self.type = kwargs['type'] - self.default_value = kwargs.get('default_value', None) - - -class VerticaLinkedService(LinkedService): - """Vertica linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: An ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(VerticaLinkedService, self).__init__(**kwargs) - self.type = 'Vertica' # type: str - self.connection_string = kwargs.get('connection_string', None) - self.pwd = kwargs.get('pwd', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class VerticaSource(TabularSource): - """A copy activity Vertica source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(VerticaSource, self).__init__(**kwargs) - self.type = 'VerticaSource' # type: str - self.query = kwargs.get('query', None) - - -class VerticaTableDataset(Dataset): - """Vertica dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The table name of the Vertica. Type: string (or Expression with resultType - string). - :type table: any - :param schema_type_properties_schema: The schema name of the Vertica. Type: string (or - Expression with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(VerticaTableDataset, self).__init__(**kwargs) - self.type = 'VerticaTable' # type: str - self.table_name = kwargs.get('table_name', None) - self.table = kwargs.get('table', None) - self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) - - -class VirtualNetworkProfile(msrest.serialization.Model): - """Virtual Network Profile. - - :param compute_subnet_id: Subnet ID used for computes in workspace. - :type compute_subnet_id: str - """ - - _attribute_map = { - 'compute_subnet_id': {'key': 'computeSubnetId', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(VirtualNetworkProfile, self).__init__(**kwargs) - self.compute_subnet_id = kwargs.get('compute_subnet_id', None) - - -class WaitActivity(ControlActivity): - """This activity suspends pipeline execution for the specified interval. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param wait_time_in_seconds: Required. Duration in seconds. - :type wait_time_in_seconds: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'wait_time_in_seconds': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'wait_time_in_seconds': {'key': 'typeProperties.waitTimeInSeconds', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(WaitActivity, self).__init__(**kwargs) - self.type = 'Wait' # type: str - self.wait_time_in_seconds = kwargs['wait_time_in_seconds'] - - -class WebActivity(ExecutionActivity): - """Web activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param method: Required. Rest API method for target endpoint. Possible values include: "GET", - "POST", "PUT", "DELETE". - :type method: str or ~azure.synapse.artifacts.v2020_12_01.models.WebActivityMethod - :param url: Required. Web activity target endpoint and path. Type: string (or Expression with - resultType string). - :type url: any - :param headers: Represents the headers that will be sent to the request. For example, to set - the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": - "application/json" }. Type: string (or Expression with resultType string). - :type headers: any - :param body: Represents the payload that will be sent to the endpoint. Required for POST/PUT - method, not allowed for GET method Type: string (or Expression with resultType string). - :type body: any - :param authentication: Authentication method used for calling the endpoint. - :type authentication: ~azure.synapse.artifacts.v2020_12_01.models.WebActivityAuthentication - :param datasets: List of datasets passed to web endpoint. - :type datasets: list[~azure.synapse.artifacts.v2020_12_01.models.DatasetReference] - :param linked_services: List of linked services passed to web endpoint. - :type linked_services: list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference] - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'method': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'method': {'key': 'typeProperties.method', 'type': 'str'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'headers': {'key': 'typeProperties.headers', 'type': 'object'}, - 'body': {'key': 'typeProperties.body', 'type': 'object'}, - 'authentication': {'key': 'typeProperties.authentication', 'type': 'WebActivityAuthentication'}, - 'datasets': {'key': 'typeProperties.datasets', 'type': '[DatasetReference]'}, - 'linked_services': {'key': 'typeProperties.linkedServices', 'type': '[LinkedServiceReference]'}, - 'connect_via': {'key': 'typeProperties.connectVia', 'type': 'IntegrationRuntimeReference'}, - } - - def __init__( - self, - **kwargs - ): - super(WebActivity, self).__init__(**kwargs) - self.type = 'WebActivity' # type: str - self.method = kwargs['method'] - self.url = kwargs['url'] - self.headers = kwargs.get('headers', None) - self.body = kwargs.get('body', None) - self.authentication = kwargs.get('authentication', None) - self.datasets = kwargs.get('datasets', None) - self.linked_services = kwargs.get('linked_services', None) - self.connect_via = kwargs.get('connect_via', None) - - -class WebActivityAuthentication(msrest.serialization.Model): - """Web activity authentication properties. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Web activity authentication (Basic/ClientCertificate/MSI). - :type type: str - :param pfx: Base64-encoded contents of a PFX file. - :type pfx: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param username: Web activity authentication user name for basic authentication. - :type username: str - :param password: Password for the PFX file or basic authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param resource: Resource for which Azure Auth token will be requested when using MSI - Authentication. - :type resource: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'pfx': {'key': 'pfx', 'type': 'SecretBase'}, - 'username': {'key': 'username', 'type': 'str'}, - 'password': {'key': 'password', 'type': 'SecretBase'}, - 'resource': {'key': 'resource', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(WebActivityAuthentication, self).__init__(**kwargs) - self.type = kwargs['type'] - self.pfx = kwargs.get('pfx', None) - self.username = kwargs.get('username', None) - self.password = kwargs.get('password', None) - self.resource = kwargs.get('resource', None) - - -class WebLinkedServiceTypeProperties(msrest.serialization.Model): - """Base definition of WebLinkedServiceTypeProperties, this typeProperties is polymorphic based on authenticationType, so not flattened in SDK models. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: WebAnonymousAuthentication, WebBasicAuthentication, WebClientCertificateAuthentication. - - All required parameters must be populated in order to send to Azure. - - :param url: Required. The URL of the web service endpoint, e.g. http://www.microsoft.com . - Type: string (or Expression with resultType string). - :type url: any - :param authentication_type: Required. Type of authentication used to connect to the web table - source.Constant filled by server. Possible values include: "Basic", "Anonymous", - "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.WebAuthenticationType - """ - - _validation = { - 'url': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'url': {'key': 'url', 'type': 'object'}, - 'authentication_type': {'key': 'authenticationType', 'type': 'str'}, - } - - _subtype_map = { - 'authentication_type': {'Anonymous': 'WebAnonymousAuthentication', 'Basic': 'WebBasicAuthentication', 'ClientCertificate': 'WebClientCertificateAuthentication'} - } - - def __init__( - self, - **kwargs - ): - super(WebLinkedServiceTypeProperties, self).__init__(**kwargs) - self.url = kwargs['url'] - self.authentication_type = None # type: Optional[str] - - -class WebAnonymousAuthentication(WebLinkedServiceTypeProperties): - """A WebLinkedService that uses anonymous authentication to communicate with an HTTP endpoint. - - All required parameters must be populated in order to send to Azure. - - :param url: Required. The URL of the web service endpoint, e.g. http://www.microsoft.com . - Type: string (or Expression with resultType string). - :type url: any - :param authentication_type: Required. Type of authentication used to connect to the web table - source.Constant filled by server. Possible values include: "Basic", "Anonymous", - "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.WebAuthenticationType - """ - - _validation = { - 'url': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'url': {'key': 'url', 'type': 'object'}, - 'authentication_type': {'key': 'authenticationType', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(WebAnonymousAuthentication, self).__init__(**kwargs) - self.authentication_type = 'Anonymous' # type: str - - -class WebBasicAuthentication(WebLinkedServiceTypeProperties): - """A WebLinkedService that uses basic authentication to communicate with an HTTP endpoint. - - All required parameters must be populated in order to send to Azure. - - :param url: Required. The URL of the web service endpoint, e.g. http://www.microsoft.com . - Type: string (or Expression with resultType string). - :type url: any - :param authentication_type: Required. Type of authentication used to connect to the web table - source.Constant filled by server. Possible values include: "Basic", "Anonymous", - "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.WebAuthenticationType - :param username: Required. User name for Basic authentication. Type: string (or Expression with - resultType string). - :type username: any - :param password: Required. The password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - """ - - _validation = { - 'url': {'required': True}, - 'authentication_type': {'required': True}, - 'username': {'required': True}, - 'password': {'required': True}, - } - - _attribute_map = { - 'url': {'key': 'url', 'type': 'object'}, - 'authentication_type': {'key': 'authenticationType', 'type': 'str'}, - 'username': {'key': 'username', 'type': 'object'}, - 'password': {'key': 'password', 'type': 'SecretBase'}, - } - - def __init__( - self, - **kwargs - ): - super(WebBasicAuthentication, self).__init__(**kwargs) - self.authentication_type = 'Basic' # type: str - self.username = kwargs['username'] - self.password = kwargs['password'] - - -class WebClientCertificateAuthentication(WebLinkedServiceTypeProperties): - """A WebLinkedService that uses client certificate based authentication to communicate with an HTTP endpoint. This scheme follows mutual authentication; the server must also provide valid credentials to the client. - - All required parameters must be populated in order to send to Azure. - - :param url: Required. The URL of the web service endpoint, e.g. http://www.microsoft.com . - Type: string (or Expression with resultType string). - :type url: any - :param authentication_type: Required. Type of authentication used to connect to the web table - source.Constant filled by server. Possible values include: "Basic", "Anonymous", - "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.WebAuthenticationType - :param pfx: Required. Base64-encoded contents of a PFX file. - :type pfx: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param password: Required. Password for the PFX file. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - """ - - _validation = { - 'url': {'required': True}, - 'authentication_type': {'required': True}, - 'pfx': {'required': True}, - 'password': {'required': True}, - } - - _attribute_map = { - 'url': {'key': 'url', 'type': 'object'}, - 'authentication_type': {'key': 'authenticationType', 'type': 'str'}, - 'pfx': {'key': 'pfx', 'type': 'SecretBase'}, - 'password': {'key': 'password', 'type': 'SecretBase'}, - } - - def __init__( - self, - **kwargs - ): - super(WebClientCertificateAuthentication, self).__init__(**kwargs) - self.authentication_type = 'ClientCertificate' # type: str - self.pfx = kwargs['pfx'] - self.password = kwargs['password'] - - -class WebHookActivity(ControlActivity): - """WebHook activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param method: Required. Rest API method for target endpoint. Possible values include: "POST". - :type method: str or ~azure.synapse.artifacts.v2020_12_01.models.WebHookActivityMethod - :param url: Required. WebHook activity target endpoint and path. Type: string (or Expression - with resultType string). - :type url: any - :param timeout: The timeout within which the webhook should be called back. If there is no - value specified, it defaults to 10 minutes. Type: string. Pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type timeout: str - :param headers: Represents the headers that will be sent to the request. For example, to set - the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": - "application/json" }. Type: string (or Expression with resultType string). - :type headers: any - :param body: Represents the payload that will be sent to the endpoint. Required for POST/PUT - method, not allowed for GET method Type: string (or Expression with resultType string). - :type body: any - :param authentication: Authentication method used for calling the endpoint. - :type authentication: ~azure.synapse.artifacts.v2020_12_01.models.WebActivityAuthentication - :param report_status_on_call_back: When set to true, statusCode, output and error in callback - request body will be consumed by activity. The activity can be marked as failed by setting - statusCode >= 400 in callback request. Default is false. Type: boolean (or Expression with - resultType boolean). - :type report_status_on_call_back: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'method': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'method': {'key': 'typeProperties.method', 'type': 'str'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'timeout': {'key': 'typeProperties.timeout', 'type': 'str'}, - 'headers': {'key': 'typeProperties.headers', 'type': 'object'}, - 'body': {'key': 'typeProperties.body', 'type': 'object'}, - 'authentication': {'key': 'typeProperties.authentication', 'type': 'WebActivityAuthentication'}, - 'report_status_on_call_back': {'key': 'typeProperties.reportStatusOnCallBack', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(WebHookActivity, self).__init__(**kwargs) - self.type = 'WebHook' # type: str - self.method = kwargs['method'] - self.url = kwargs['url'] - self.timeout = kwargs.get('timeout', None) - self.headers = kwargs.get('headers', None) - self.body = kwargs.get('body', None) - self.authentication = kwargs.get('authentication', None) - self.report_status_on_call_back = kwargs.get('report_status_on_call_back', None) - - -class WebLinkedService(LinkedService): - """Web linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param type_properties: Required. Web linked service properties. - :type type_properties: - ~azure.synapse.artifacts.v2020_12_01.models.WebLinkedServiceTypeProperties - """ - - _validation = { - 'type': {'required': True}, - 'type_properties': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'type_properties': {'key': 'typeProperties', 'type': 'WebLinkedServiceTypeProperties'}, - } - - def __init__( - self, - **kwargs - ): - super(WebLinkedService, self).__init__(**kwargs) - self.type = 'Web' # type: str - self.type_properties = kwargs['type_properties'] - - -class WebSource(CopySource): - """A copy activity source for web page table. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(WebSource, self).__init__(**kwargs) - self.type = 'WebSource' # type: str - self.additional_columns = kwargs.get('additional_columns', None) - - -class WebTableDataset(Dataset): - """The dataset points to a HTML table in the web page. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param index: Required. The zero-based index of the table in the web page. Type: integer (or - Expression with resultType integer), minimum: 0. - :type index: any - :param path: The relative URL to the web page from the linked service URL. Type: string (or - Expression with resultType string). - :type path: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'index': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'index': {'key': 'typeProperties.index', 'type': 'object'}, - 'path': {'key': 'typeProperties.path', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(WebTableDataset, self).__init__(**kwargs) - self.type = 'WebTable' # type: str - self.index = kwargs['index'] - self.path = kwargs.get('path', None) - - -class Workspace(TrackedResource): - """A workspace. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :param tags: A set of tags. Resource tags. - :type tags: dict[str, str] - :param location: Required. The geo-location where the resource lives. - :type location: str - :param identity: Identity of the workspace. - :type identity: ~azure.synapse.artifacts.v2020_12_01.models.ManagedIdentity - :param default_data_lake_storage: Workspace default data lake storage account details. - :type default_data_lake_storage: - ~azure.synapse.artifacts.v2020_12_01.models.DataLakeStorageAccountDetails - :param sql_administrator_login_password: SQL administrator login password. - :type sql_administrator_login_password: str - :param managed_resource_group_name: Workspace managed resource group. The resource group name - uniquely identifies the resource group within the user subscriptionId. The resource group name - must be no longer than 90 characters long, and must be alphanumeric characters - (Char.IsLetterOrDigit()) and '-', '_', '(', ')' and'.'. Note that the name cannot end with '.'. - :type managed_resource_group_name: str - :ivar provisioning_state: Resource provisioning state. - :vartype provisioning_state: str - :param sql_administrator_login: Login for workspace SQL active directory administrator. - :type sql_administrator_login: str - :param virtual_network_profile: Virtual Network profile. - :type virtual_network_profile: - ~azure.synapse.artifacts.v2020_12_01.models.VirtualNetworkProfile - :param connectivity_endpoints: Connectivity endpoints. - :type connectivity_endpoints: dict[str, str] - :param managed_virtual_network: Setting this to 'default' will ensure that all compute for this - workspace is in a virtual network managed on behalf of the user. - :type managed_virtual_network: str - :param private_endpoint_connections: Private endpoint connections to the workspace. - :type private_endpoint_connections: - list[~azure.synapse.artifacts.v2020_12_01.models.PrivateEndpointConnection] - :param encryption: The encryption details of the workspace. - :type encryption: ~azure.synapse.artifacts.v2020_12_01.models.EncryptionDetails - :ivar workspace_uid: The workspace unique identifier. - :vartype workspace_uid: str - :ivar extra_properties: Workspace level configs and feature flags. - :vartype extra_properties: dict[str, any] - :param managed_virtual_network_settings: Managed Virtual Network Settings. - :type managed_virtual_network_settings: - ~azure.synapse.artifacts.v2020_12_01.models.ManagedVirtualNetworkSettings - :param workspace_repository_configuration: Git integration settings. - :type workspace_repository_configuration: - ~azure.synapse.artifacts.v2020_12_01.models.WorkspaceRepositoryConfiguration - :param purview_configuration: Purview Configuration. - :type purview_configuration: ~azure.synapse.artifacts.v2020_12_01.models.PurviewConfiguration - :ivar adla_resource_id: The ADLA resource ID. - :vartype adla_resource_id: str - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'location': {'required': True}, - 'provisioning_state': {'readonly': True}, - 'workspace_uid': {'readonly': True}, - 'extra_properties': {'readonly': True}, - 'adla_resource_id': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'tags': {'key': 'tags', 'type': '{str}'}, - 'location': {'key': 'location', 'type': 'str'}, - 'identity': {'key': 'identity', 'type': 'ManagedIdentity'}, - 'default_data_lake_storage': {'key': 'properties.defaultDataLakeStorage', 'type': 'DataLakeStorageAccountDetails'}, - 'sql_administrator_login_password': {'key': 'properties.sqlAdministratorLoginPassword', 'type': 'str'}, - 'managed_resource_group_name': {'key': 'properties.managedResourceGroupName', 'type': 'str'}, - 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, - 'sql_administrator_login': {'key': 'properties.sqlAdministratorLogin', 'type': 'str'}, - 'virtual_network_profile': {'key': 'properties.virtualNetworkProfile', 'type': 'VirtualNetworkProfile'}, - 'connectivity_endpoints': {'key': 'properties.connectivityEndpoints', 'type': '{str}'}, - 'managed_virtual_network': {'key': 'properties.managedVirtualNetwork', 'type': 'str'}, - 'private_endpoint_connections': {'key': 'properties.privateEndpointConnections', 'type': '[PrivateEndpointConnection]'}, - 'encryption': {'key': 'properties.encryption', 'type': 'EncryptionDetails'}, - 'workspace_uid': {'key': 'properties.workspaceUID', 'type': 'str'}, - 'extra_properties': {'key': 'properties.extraProperties', 'type': '{object}'}, - 'managed_virtual_network_settings': {'key': 'properties.managedVirtualNetworkSettings', 'type': 'ManagedVirtualNetworkSettings'}, - 'workspace_repository_configuration': {'key': 'properties.workspaceRepositoryConfiguration', 'type': 'WorkspaceRepositoryConfiguration'}, - 'purview_configuration': {'key': 'properties.purviewConfiguration', 'type': 'PurviewConfiguration'}, - 'adla_resource_id': {'key': 'properties.adlaResourceId', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(Workspace, self).__init__(**kwargs) - self.identity = kwargs.get('identity', None) - self.default_data_lake_storage = kwargs.get('default_data_lake_storage', None) - self.sql_administrator_login_password = kwargs.get('sql_administrator_login_password', None) - self.managed_resource_group_name = kwargs.get('managed_resource_group_name', None) - self.provisioning_state = None - self.sql_administrator_login = kwargs.get('sql_administrator_login', None) - self.virtual_network_profile = kwargs.get('virtual_network_profile', None) - self.connectivity_endpoints = kwargs.get('connectivity_endpoints', None) - self.managed_virtual_network = kwargs.get('managed_virtual_network', None) - self.private_endpoint_connections = kwargs.get('private_endpoint_connections', None) - self.encryption = kwargs.get('encryption', None) - self.workspace_uid = None - self.extra_properties = None - self.managed_virtual_network_settings = kwargs.get('managed_virtual_network_settings', None) - self.workspace_repository_configuration = kwargs.get('workspace_repository_configuration', None) - self.purview_configuration = kwargs.get('purview_configuration', None) - self.adla_resource_id = None - - -class WorkspaceIdentity(msrest.serialization.Model): - """Identity properties of the workspace resource. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar type: The identity type. Currently the only supported type is 'SystemAssigned'. Has - constant value: "SystemAssigned". - :vartype type: str - :ivar principal_id: The principal id of the identity. - :vartype principal_id: str - :ivar tenant_id: The client tenant id of the identity. - :vartype tenant_id: str - """ - - _validation = { - 'type': {'required': True, 'constant': True}, - 'principal_id': {'readonly': True}, - 'tenant_id': {'readonly': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'principal_id': {'key': 'principalId', 'type': 'str'}, - 'tenant_id': {'key': 'tenantId', 'type': 'str'}, - } - - type = "SystemAssigned" - - def __init__( - self, - **kwargs - ): - super(WorkspaceIdentity, self).__init__(**kwargs) - self.principal_id = None - self.tenant_id = None - - -class WorkspaceKeyDetails(msrest.serialization.Model): - """Details of the customer managed key associated with the workspace. - - :param name: Workspace Key sub-resource name. - :type name: str - :param key_vault_url: Workspace Key sub-resource key vault url. - :type key_vault_url: str - """ - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'key_vault_url': {'key': 'keyVaultUrl', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(WorkspaceKeyDetails, self).__init__(**kwargs) - self.name = kwargs.get('name', None) - self.key_vault_url = kwargs.get('key_vault_url', None) - - -class WorkspaceRepositoryConfiguration(msrest.serialization.Model): - """Git integration settings. - - :param type: Type of workspace repositoryID configuration. Example WorkspaceVSTSConfiguration, - WorkspaceGitHubConfiguration. - :type type: str - :param host_name: GitHub Enterprise host name. For example: https://github.mydomain.com. - :type host_name: str - :param account_name: Account name. - :type account_name: str - :param project_name: VSTS project name. - :type project_name: str - :param repository_name: Repository name. - :type repository_name: str - :param collaboration_branch: Collaboration branch. - :type collaboration_branch: str - :param root_folder: Root folder to use in the repository. - :type root_folder: str - :param last_commit_id: The last commit ID. - :type last_commit_id: str - :param tenant_id: The VSTS tenant ID. - :type tenant_id: str - :param client_id: GitHub bring your own app client id. - :type client_id: str - :param client_secret: GitHub bring your own app client secret information. - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.GitHubClientSecret - """ - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'host_name': {'key': 'hostName', 'type': 'str'}, - 'account_name': {'key': 'accountName', 'type': 'str'}, - 'project_name': {'key': 'projectName', 'type': 'str'}, - 'repository_name': {'key': 'repositoryName', 'type': 'str'}, - 'collaboration_branch': {'key': 'collaborationBranch', 'type': 'str'}, - 'root_folder': {'key': 'rootFolder', 'type': 'str'}, - 'last_commit_id': {'key': 'lastCommitId', 'type': 'str'}, - 'tenant_id': {'key': 'tenantId', 'type': 'str'}, - 'client_id': {'key': 'clientId', 'type': 'str'}, - 'client_secret': {'key': 'clientSecret', 'type': 'GitHubClientSecret'}, - } - - def __init__( - self, - **kwargs - ): - super(WorkspaceRepositoryConfiguration, self).__init__(**kwargs) - self.type = kwargs.get('type', None) - self.host_name = kwargs.get('host_name', None) - self.account_name = kwargs.get('account_name', None) - self.project_name = kwargs.get('project_name', None) - self.repository_name = kwargs.get('repository_name', None) - self.collaboration_branch = kwargs.get('collaboration_branch', None) - self.root_folder = kwargs.get('root_folder', None) - self.last_commit_id = kwargs.get('last_commit_id', None) - self.tenant_id = kwargs.get('tenant_id', None) - self.client_id = kwargs.get('client_id', None) - self.client_secret = kwargs.get('client_secret', None) - - -class WorkspaceUpdateParameters(msrest.serialization.Model): - """Parameters for updating a workspace resource. - - :param tags: A set of tags. The resource tags. - :type tags: dict[str, str] - :param identity: Managed service identity of the workspace. - :type identity: ~azure.synapse.artifacts.v2020_12_01.models.WorkspaceIdentity - """ - - _attribute_map = { - 'tags': {'key': 'tags', 'type': '{str}'}, - 'identity': {'key': 'identity', 'type': 'WorkspaceIdentity'}, - } - - def __init__( - self, - **kwargs - ): - super(WorkspaceUpdateParameters, self).__init__(**kwargs) - self.tags = kwargs.get('tags', None) - self.identity = kwargs.get('identity', None) - - -class XeroLinkedService(LinkedService): - """Xero Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_properties: Properties used to connect to Xero. It is mutually exclusive with - any other properties in the linked service. Type: object. - :type connection_properties: any - :param host: Required. The endpoint of the Xero server. (i.e. api.xero.com). - :type host: any - :param consumer_key: The consumer key associated with the Xero application. - :type consumer_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param private_key: The private key from the .pem file that was generated for your Xero private - application. You must include all the text from the .pem file, including the Unix line endings( - ). - :type private_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_properties': {'key': 'typeProperties.connectionProperties', 'type': 'object'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'consumer_key': {'key': 'typeProperties.consumerKey', 'type': 'SecretBase'}, - 'private_key': {'key': 'typeProperties.privateKey', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(XeroLinkedService, self).__init__(**kwargs) - self.type = 'Xero' # type: str - self.connection_properties = kwargs.get('connection_properties', None) - self.host = kwargs['host'] - self.consumer_key = kwargs.get('consumer_key', None) - self.private_key = kwargs.get('private_key', None) - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.use_host_verification = kwargs.get('use_host_verification', None) - self.use_peer_verification = kwargs.get('use_peer_verification', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class XeroObjectDataset(Dataset): - """Xero Service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(XeroObjectDataset, self).__init__(**kwargs) - self.type = 'XeroObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class XeroSource(TabularSource): - """A copy activity Xero Service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(XeroSource, self).__init__(**kwargs) - self.type = 'XeroSource' # type: str - self.query = kwargs.get('query', None) - - -class XmlDataset(Dataset): - """Xml dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param location: The location of the json data storage. - :type location: ~azure.synapse.artifacts.v2020_12_01.models.DatasetLocation - :param encoding_name: The code page name of the preferred encoding. If not specified, the - default value is UTF-8, unless BOM denotes another Unicode encoding. Refer to the name column - of the table in the following link to set supported values: - https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string (or Expression with - resultType string). - :type encoding_name: any - :param null_value: The null value string. Type: string (or Expression with resultType string). - :type null_value: any - :param compression: The data compression method used for the json dataset. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'location': {'key': 'typeProperties.location', 'type': 'DatasetLocation'}, - 'encoding_name': {'key': 'typeProperties.encodingName', 'type': 'object'}, - 'null_value': {'key': 'typeProperties.nullValue', 'type': 'object'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - } - - def __init__( - self, - **kwargs - ): - super(XmlDataset, self).__init__(**kwargs) - self.type = 'Xml' # type: str - self.location = kwargs.get('location', None) - self.encoding_name = kwargs.get('encoding_name', None) - self.null_value = kwargs.get('null_value', None) - self.compression = kwargs.get('compression', None) - - -class XmlReadSettings(FormatReadSettings): - """Xml read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param compression_properties: Compression settings. - :type compression_properties: - ~azure.synapse.artifacts.v2020_12_01.models.CompressionReadSettings - :param validation_mode: Indicates what validation method is used when reading the xml files. - Allowed values: 'none', 'xsd', or 'dtd'. Type: string (or Expression with resultType string). - :type validation_mode: any - :param detect_data_type: Indicates whether type detection is enabled when reading the xml - files. Type: boolean (or Expression with resultType boolean). - :type detect_data_type: any - :param namespaces: Indicates whether namespace is enabled when reading the xml files. Type: - boolean (or Expression with resultType boolean). - :type namespaces: any - :param namespace_prefixes: Namespace uri to prefix mappings to override the prefixes in column - names when namespace is enabled, if no prefix is defined for a namespace uri, the prefix of xml - element/attribute name in the xml data file will be used. Example: - "{"http://www.example.com/xml":"prefix"}" Type: object (or Expression with resultType object). - :type namespace_prefixes: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'compression_properties': {'key': 'compressionProperties', 'type': 'CompressionReadSettings'}, - 'validation_mode': {'key': 'validationMode', 'type': 'object'}, - 'detect_data_type': {'key': 'detectDataType', 'type': 'object'}, - 'namespaces': {'key': 'namespaces', 'type': 'object'}, - 'namespace_prefixes': {'key': 'namespacePrefixes', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(XmlReadSettings, self).__init__(**kwargs) - self.type = 'XmlReadSettings' # type: str - self.compression_properties = kwargs.get('compression_properties', None) - self.validation_mode = kwargs.get('validation_mode', None) - self.detect_data_type = kwargs.get('detect_data_type', None) - self.namespaces = kwargs.get('namespaces', None) - self.namespace_prefixes = kwargs.get('namespace_prefixes', None) - - -class XmlSource(CopySource): - """A copy activity Xml source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Xml store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - :param format_settings: Xml format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.XmlReadSettings - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'format_settings': {'key': 'formatSettings', 'type': 'XmlReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(XmlSource, self).__init__(**kwargs) - self.type = 'XmlSource' # type: str - self.store_settings = kwargs.get('store_settings', None) - self.format_settings = kwargs.get('format_settings', None) - self.additional_columns = kwargs.get('additional_columns', None) - - -class ZipDeflateReadSettings(CompressionReadSettings): - """The ZipDeflate compression read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The Compression setting type.Constant filled by server. - :type type: str - :param preserve_zip_file_name_as_folder: Preserve the zip file name as folder path. Type: - boolean (or Expression with resultType boolean). - :type preserve_zip_file_name_as_folder: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'preserve_zip_file_name_as_folder': {'key': 'preserveZipFileNameAsFolder', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ZipDeflateReadSettings, self).__init__(**kwargs) - self.type = 'ZipDeflateReadSettings' # type: str - self.preserve_zip_file_name_as_folder = kwargs.get('preserve_zip_file_name_as_folder', None) - - -class ZohoLinkedService(LinkedService): - """Zoho server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_properties: Properties used to connect to Zoho. It is mutually exclusive with - any other properties in the linked service. Type: object. - :type connection_properties: any - :param endpoint: Required. The endpoint of the Zoho server. (i.e. crm.zoho.com/crm/private). - :type endpoint: any - :param access_token: The access token for Zoho authentication. - :type access_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'endpoint': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_properties': {'key': 'typeProperties.connectionProperties', 'type': 'object'}, - 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, - 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ZohoLinkedService, self).__init__(**kwargs) - self.type = 'Zoho' # type: str - self.connection_properties = kwargs.get('connection_properties', None) - self.endpoint = kwargs['endpoint'] - self.access_token = kwargs.get('access_token', None) - self.use_encrypted_endpoints = kwargs.get('use_encrypted_endpoints', None) - self.use_host_verification = kwargs.get('use_host_verification', None) - self.use_peer_verification = kwargs.get('use_peer_verification', None) - self.encrypted_credential = kwargs.get('encrypted_credential', None) - - -class ZohoObjectDataset(Dataset): - """Zoho server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ZohoObjectDataset, self).__init__(**kwargs) - self.type = 'ZohoObject' # type: str - self.table_name = kwargs.get('table_name', None) - - -class ZohoSource(TabularSource): - """A copy activity Zoho server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ZohoSource, self).__init__(**kwargs) - self.type = 'ZohoSource' # type: str - self.query = kwargs.get('query', None) diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/models/_models_py3.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/models/_models_py3.py deleted file mode 100644 index 24205e278893..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/models/_models_py3.py +++ /dev/null @@ -1,41346 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -import datetime -from typing import Any, Dict, List, Optional, Union - -from azure.core.exceptions import HttpResponseError -import msrest.serialization - -from ._artifacts_client_enums import * - - -class Activity(msrest.serialization.Model): - """A pipeline activity. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: ControlActivity, ExecutionActivity, SqlPoolStoredProcedureActivity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - } - - _subtype_map = { - 'type': {'Container': 'ControlActivity', 'Execution': 'ExecutionActivity', 'SqlPoolStoredProcedure': 'SqlPoolStoredProcedureActivity'} - } - - def __init__( - self, - *, - name: str, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - **kwargs - ): - super(Activity, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.name = name - self.type = 'Activity' # type: str - self.description = description - self.depends_on = depends_on - self.user_properties = user_properties - - -class ActivityDependency(msrest.serialization.Model): - """Activity dependency information. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param activity: Required. Activity name. - :type activity: str - :param dependency_conditions: Required. Match-Condition for the dependency. - :type dependency_conditions: list[str or - ~azure.synapse.artifacts.v2020_12_01.models.DependencyCondition] - """ - - _validation = { - 'activity': {'required': True}, - 'dependency_conditions': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'activity': {'key': 'activity', 'type': 'str'}, - 'dependency_conditions': {'key': 'dependencyConditions', 'type': '[str]'}, - } - - def __init__( - self, - *, - activity: str, - dependency_conditions: List[Union[str, "DependencyCondition"]], - additional_properties: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(ActivityDependency, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.activity = activity - self.dependency_conditions = dependency_conditions - - -class ActivityPolicy(msrest.serialization.Model): - """Execution policy for an activity. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param timeout: Specifies the timeout for the activity to run. The default timeout is 7 days. - Type: string (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type timeout: any - :param retry: Maximum ordinary retry attempts. Default is 0. Type: integer (or Expression with - resultType integer), minimum: 0. - :type retry: any - :param retry_interval_in_seconds: Interval between each retry attempt (in seconds). The default - is 30 sec. - :type retry_interval_in_seconds: int - :param secure_input: When set to true, Input from activity is considered as secure and will not - be logged to monitoring. - :type secure_input: bool - :param secure_output: When set to true, Output from activity is considered as secure and will - not be logged to monitoring. - :type secure_output: bool - """ - - _validation = { - 'retry_interval_in_seconds': {'maximum': 86400, 'minimum': 30}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'timeout': {'key': 'timeout', 'type': 'object'}, - 'retry': {'key': 'retry', 'type': 'object'}, - 'retry_interval_in_seconds': {'key': 'retryIntervalInSeconds', 'type': 'int'}, - 'secure_input': {'key': 'secureInput', 'type': 'bool'}, - 'secure_output': {'key': 'secureOutput', 'type': 'bool'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - timeout: Optional[Any] = None, - retry: Optional[Any] = None, - retry_interval_in_seconds: Optional[int] = None, - secure_input: Optional[bool] = None, - secure_output: Optional[bool] = None, - **kwargs - ): - super(ActivityPolicy, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.timeout = timeout - self.retry = retry - self.retry_interval_in_seconds = retry_interval_in_seconds - self.secure_input = secure_input - self.secure_output = secure_output - - -class ActivityRun(msrest.serialization.Model): - """Information about an activity run in a pipeline. - - Variables are only populated by the server, and will be ignored when sending a request. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :ivar pipeline_name: The name of the pipeline. - :vartype pipeline_name: str - :ivar pipeline_run_id: The id of the pipeline run. - :vartype pipeline_run_id: str - :ivar activity_name: The name of the activity. - :vartype activity_name: str - :ivar activity_type: The type of the activity. - :vartype activity_type: str - :ivar activity_run_id: The id of the activity run. - :vartype activity_run_id: str - :ivar linked_service_name: The name of the compute linked service. - :vartype linked_service_name: str - :ivar status: The status of the activity run. - :vartype status: str - :ivar activity_run_start: The start time of the activity run in 'ISO 8601' format. - :vartype activity_run_start: ~datetime.datetime - :ivar activity_run_end: The end time of the activity run in 'ISO 8601' format. - :vartype activity_run_end: ~datetime.datetime - :ivar duration_in_ms: The duration of the activity run. - :vartype duration_in_ms: int - :ivar input: The input for the activity. - :vartype input: any - :ivar output: The output for the activity. - :vartype output: any - :ivar error: The error if any from the activity run. - :vartype error: any - """ - - _validation = { - 'pipeline_name': {'readonly': True}, - 'pipeline_run_id': {'readonly': True}, - 'activity_name': {'readonly': True}, - 'activity_type': {'readonly': True}, - 'activity_run_id': {'readonly': True}, - 'linked_service_name': {'readonly': True}, - 'status': {'readonly': True}, - 'activity_run_start': {'readonly': True}, - 'activity_run_end': {'readonly': True}, - 'duration_in_ms': {'readonly': True}, - 'input': {'readonly': True}, - 'output': {'readonly': True}, - 'error': {'readonly': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'pipeline_name': {'key': 'pipelineName', 'type': 'str'}, - 'pipeline_run_id': {'key': 'pipelineRunId', 'type': 'str'}, - 'activity_name': {'key': 'activityName', 'type': 'str'}, - 'activity_type': {'key': 'activityType', 'type': 'str'}, - 'activity_run_id': {'key': 'activityRunId', 'type': 'str'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'str'}, - 'status': {'key': 'status', 'type': 'str'}, - 'activity_run_start': {'key': 'activityRunStart', 'type': 'iso-8601'}, - 'activity_run_end': {'key': 'activityRunEnd', 'type': 'iso-8601'}, - 'duration_in_ms': {'key': 'durationInMs', 'type': 'int'}, - 'input': {'key': 'input', 'type': 'object'}, - 'output': {'key': 'output', 'type': 'object'}, - 'error': {'key': 'error', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(ActivityRun, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.pipeline_name = None - self.pipeline_run_id = None - self.activity_name = None - self.activity_type = None - self.activity_run_id = None - self.linked_service_name = None - self.status = None - self.activity_run_start = None - self.activity_run_end = None - self.duration_in_ms = None - self.input = None - self.output = None - self.error = None - - -class ActivityRunsQueryResponse(msrest.serialization.Model): - """A list activity runs. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of activity runs. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityRun] - :param continuation_token: The continuation token for getting the next page of results, if any - remaining results exist, null otherwise. - :type continuation_token: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[ActivityRun]'}, - 'continuation_token': {'key': 'continuationToken', 'type': 'str'}, - } - - def __init__( - self, - *, - value: List["ActivityRun"], - continuation_token: Optional[str] = None, - **kwargs - ): - super(ActivityRunsQueryResponse, self).__init__(**kwargs) - self.value = value - self.continuation_token = continuation_token - - -class AddDataFlowToDebugSessionResponse(msrest.serialization.Model): - """Response body structure for starting data flow debug session. - - :param job_version: The ID of data flow debug job version. - :type job_version: str - """ - - _attribute_map = { - 'job_version': {'key': 'jobVersion', 'type': 'str'}, - } - - def __init__( - self, - *, - job_version: Optional[str] = None, - **kwargs - ): - super(AddDataFlowToDebugSessionResponse, self).__init__(**kwargs) - self.job_version = job_version - - -class AdditionalColumns(msrest.serialization.Model): - """Specify the column name and value of additional columns. - - :param name: Additional column name. Type: string (or Expression with resultType string). - :type name: any - :param value: Additional column value. Type: string (or Expression with resultType string). - :type value: any - """ - - _attribute_map = { - 'name': {'key': 'name', 'type': 'object'}, - 'value': {'key': 'value', 'type': 'object'}, - } - - def __init__( - self, - *, - name: Optional[Any] = None, - value: Optional[Any] = None, - **kwargs - ): - super(AdditionalColumns, self).__init__(**kwargs) - self.name = name - self.value = value - - -class LinkedService(msrest.serialization.Model): - """The Azure Synapse nested object which contains the information and credential which can be used to connect with related store or compute resource. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonMWSLinkedService, AmazonRedshiftLinkedService, AmazonS3LinkedService, AzureBatchLinkedService, AzureBlobFSLinkedService, AzureBlobStorageLinkedService, AzureDataExplorerLinkedService, AzureDataLakeAnalyticsLinkedService, AzureDataLakeStoreLinkedService, AzureDatabricksLinkedService, AzureDatabricksDeltaLakeLinkedService, AzureFileStorageLinkedService, AzureFunctionLinkedService, AzureKeyVaultLinkedService, AzureMLLinkedService, AzureMLServiceLinkedService, AzureMariaDBLinkedService, AzureMySqlLinkedService, AzurePostgreSqlLinkedService, AzureSearchLinkedService, AzureSqlDWLinkedService, AzureSqlDatabaseLinkedService, AzureSqlMILinkedService, AzureStorageLinkedService, AzureTableStorageLinkedService, CassandraLinkedService, CommonDataServiceForAppsLinkedService, ConcurLinkedService, CosmosDbLinkedService, CosmosDbMongoDbApiLinkedService, CouchbaseLinkedService, CustomDataSourceLinkedService, Db2LinkedService, DrillLinkedService, DynamicsLinkedService, DynamicsAXLinkedService, DynamicsCrmLinkedService, EloquaLinkedService, FileServerLinkedService, FtpServerLinkedService, GoogleAdWordsLinkedService, GoogleBigQueryLinkedService, GoogleCloudStorageLinkedService, GreenplumLinkedService, HBaseLinkedService, HDInsightLinkedService, HDInsightOnDemandLinkedService, HdfsLinkedService, HiveLinkedService, HttpLinkedService, HubspotLinkedService, ImpalaLinkedService, InformixLinkedService, JiraLinkedService, MagentoLinkedService, MariaDBLinkedService, MarketoLinkedService, MicrosoftAccessLinkedService, MongoDbLinkedService, MongoDbAtlasLinkedService, MongoDbV2LinkedService, MySqlLinkedService, NetezzaLinkedService, ODataLinkedService, OdbcLinkedService, Office365LinkedService, OracleLinkedService, OracleServiceCloudLinkedService, PaypalLinkedService, PhoenixLinkedService, PostgreSqlLinkedService, PrestoLinkedService, QuickBooksLinkedService, ResponsysLinkedService, RestServiceLinkedService, SalesforceLinkedService, SalesforceMarketingCloudLinkedService, SalesforceServiceCloudLinkedService, SapBWLinkedService, SapCloudForCustomerLinkedService, SapEccLinkedService, SapHanaLinkedService, SapOpenHubLinkedService, SapTableLinkedService, ServiceNowLinkedService, SftpServerLinkedService, SharePointOnlineListLinkedService, ShopifyLinkedService, SnowflakeLinkedService, SparkLinkedService, SqlServerLinkedService, SquareLinkedService, SybaseLinkedService, TeradataLinkedService, VerticaLinkedService, WebLinkedService, XeroLinkedService, ZohoLinkedService. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - } - - _subtype_map = { - 'type': {'AmazonMWS': 'AmazonMWSLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureBlobFS': 'AzureBlobFSLinkedService', 'AzureBlobStorage': 'AzureBlobStorageLinkedService', 'AzureDataExplorer': 'AzureDataExplorerLinkedService', 'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'AzureDatabricks': 'AzureDatabricksLinkedService', 'AzureDatabricksDeltaLake': 'AzureDatabricksDeltaLakeLinkedService', 'AzureFileStorage': 'AzureFileStorageLinkedService', 'AzureFunction': 'AzureFunctionLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureML': 'AzureMLLinkedService', 'AzureMLService': 'AzureMLServiceLinkedService', 'AzureMariaDB': 'AzureMariaDBLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'AzureSqlMI': 'AzureSqlMILinkedService', 'AzureStorage': 'AzureStorageLinkedService', 'AzureTableStorage': 'AzureTableStorageLinkedService', 'Cassandra': 'CassandraLinkedService', 'CommonDataServiceForApps': 'CommonDataServiceForAppsLinkedService', 'Concur': 'ConcurLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'CosmosDbMongoDbApi': 'CosmosDbMongoDbApiLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'Db2': 'Db2LinkedService', 'Drill': 'DrillLinkedService', 'Dynamics': 'DynamicsLinkedService', 'DynamicsAX': 'DynamicsAXLinkedService', 'DynamicsCrm': 'DynamicsCrmLinkedService', 'Eloqua': 'EloquaLinkedService', 'FileServer': 'FileServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'GoogleAdWords': 'GoogleAdWordsLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'GoogleCloudStorage': 'GoogleCloudStorageLinkedService', 'Greenplum': 'GreenplumLinkedService', 'HBase': 'HBaseLinkedService', 'HDInsight': 'HDInsightLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'Hdfs': 'HdfsLinkedService', 'Hive': 'HiveLinkedService', 'HttpServer': 'HttpLinkedService', 'Hubspot': 'HubspotLinkedService', 'Impala': 'ImpalaLinkedService', 'Informix': 'InformixLinkedService', 'Jira': 'JiraLinkedService', 'Magento': 'MagentoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Marketo': 'MarketoLinkedService', 'MicrosoftAccess': 'MicrosoftAccessLinkedService', 'MongoDb': 'MongoDbLinkedService', 'MongoDbAtlas': 'MongoDbAtlasLinkedService', 'MongoDbV2': 'MongoDbV2LinkedService', 'MySql': 'MySqlLinkedService', 'Netezza': 'NetezzaLinkedService', 'OData': 'ODataLinkedService', 'Odbc': 'OdbcLinkedService', 'Office365': 'Office365LinkedService', 'Oracle': 'OracleLinkedService', 'OracleServiceCloud': 'OracleServiceCloudLinkedService', 'Paypal': 'PaypalLinkedService', 'Phoenix': 'PhoenixLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'Presto': 'PrestoLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Responsys': 'ResponsysLinkedService', 'RestService': 'RestServiceLinkedService', 'Salesforce': 'SalesforceLinkedService', 'SalesforceMarketingCloud': 'SalesforceMarketingCloudLinkedService', 'SalesforceServiceCloud': 'SalesforceServiceCloudLinkedService', 'SapBW': 'SapBWLinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'SapEcc': 'SapEccLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapOpenHub': 'SapOpenHubLinkedService', 'SapTable': 'SapTableLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'Sftp': 'SftpServerLinkedService', 'SharePointOnlineList': 'SharePointOnlineListLinkedService', 'Shopify': 'ShopifyLinkedService', 'Snowflake': 'SnowflakeLinkedService', 'Spark': 'SparkLinkedService', 'SqlServer': 'SqlServerLinkedService', 'Square': 'SquareLinkedService', 'Sybase': 'SybaseLinkedService', 'Teradata': 'TeradataLinkedService', 'Vertica': 'VerticaLinkedService', 'Web': 'WebLinkedService', 'Xero': 'XeroLinkedService', 'Zoho': 'ZohoLinkedService'} - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - **kwargs - ): - super(LinkedService, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = 'LinkedService' # type: str - self.connect_via = connect_via - self.description = description - self.parameters = parameters - self.annotations = annotations - - -class AmazonMWSLinkedService(LinkedService): - """Amazon Marketplace Web Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param endpoint: Required. The endpoint of the Amazon MWS server, (i.e. - mws.amazonservices.com). - :type endpoint: any - :param marketplace_id: Required. The Amazon Marketplace ID you want to retrieve data from. To - retrieve data from multiple Marketplace IDs, separate them with a comma (,). (i.e. - A2EUQ1WTGCTBG2). - :type marketplace_id: any - :param seller_id: Required. The Amazon seller ID. - :type seller_id: any - :param mws_auth_token: The Amazon MWS authentication token. - :type mws_auth_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param access_key_id: Required. The access key id used to access data. - :type access_key_id: any - :param secret_key: The secret key used to access data. - :type secret_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'endpoint': {'required': True}, - 'marketplace_id': {'required': True}, - 'seller_id': {'required': True}, - 'access_key_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, - 'marketplace_id': {'key': 'typeProperties.marketplaceID', 'type': 'object'}, - 'seller_id': {'key': 'typeProperties.sellerID', 'type': 'object'}, - 'mws_auth_token': {'key': 'typeProperties.mwsAuthToken', 'type': 'SecretBase'}, - 'access_key_id': {'key': 'typeProperties.accessKeyId', 'type': 'object'}, - 'secret_key': {'key': 'typeProperties.secretKey', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - endpoint: Any, - marketplace_id: Any, - seller_id: Any, - access_key_id: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - mws_auth_token: Optional["SecretBase"] = None, - secret_key: Optional["SecretBase"] = None, - use_encrypted_endpoints: Optional[Any] = None, - use_host_verification: Optional[Any] = None, - use_peer_verification: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AmazonMWSLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AmazonMWS' # type: str - self.endpoint = endpoint - self.marketplace_id = marketplace_id - self.seller_id = seller_id - self.mws_auth_token = mws_auth_token - self.access_key_id = access_key_id - self.secret_key = secret_key - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential - - -class Dataset(msrest.serialization.Model): - """The Azure Data Factory nested object which identifies data within different data stores, such as tables, files, folders, and documents. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonMWSObjectDataset, AmazonRedshiftTableDataset, AmazonS3Dataset, AvroDataset, AzureBlobDataset, AzureBlobFSDataset, AzureDataExplorerTableDataset, AzureDataLakeStoreDataset, AzureDatabricksDeltaLakeDataset, AzureMariaDBTableDataset, AzureMySqlTableDataset, AzurePostgreSqlTableDataset, AzureSearchIndexDataset, AzureSqlDWTableDataset, AzureSqlMITableDataset, AzureSqlTableDataset, AzureTableDataset, BinaryDataset, CassandraTableDataset, CommonDataServiceForAppsEntityDataset, ConcurObjectDataset, CosmosDbMongoDbApiCollectionDataset, CosmosDbSqlApiCollectionDataset, CouchbaseTableDataset, CustomDataset, Db2TableDataset, DelimitedTextDataset, DocumentDbCollectionDataset, DrillTableDataset, DynamicsAXResourceDataset, DynamicsCrmEntityDataset, DynamicsEntityDataset, EloquaObjectDataset, ExcelDataset, FileShareDataset, GoogleAdWordsObjectDataset, GoogleBigQueryObjectDataset, GreenplumTableDataset, HBaseObjectDataset, HiveObjectDataset, HttpDataset, HubspotObjectDataset, ImpalaObjectDataset, InformixTableDataset, JiraObjectDataset, JsonDataset, MagentoObjectDataset, MariaDBTableDataset, MarketoObjectDataset, MicrosoftAccessTableDataset, MongoDbAtlasCollectionDataset, MongoDbCollectionDataset, MongoDbV2CollectionDataset, MySqlTableDataset, NetezzaTableDataset, ODataResourceDataset, OdbcTableDataset, Office365Dataset, OracleServiceCloudObjectDataset, OracleTableDataset, OrcDataset, ParquetDataset, PaypalObjectDataset, PhoenixObjectDataset, PostgreSqlTableDataset, PrestoObjectDataset, QuickBooksObjectDataset, RelationalTableDataset, ResponsysObjectDataset, RestResourceDataset, SalesforceMarketingCloudObjectDataset, SalesforceObjectDataset, SalesforceServiceCloudObjectDataset, SapBwCubeDataset, SapCloudForCustomerResourceDataset, SapEccResourceDataset, SapHanaTableDataset, SapOpenHubTableDataset, SapTableResourceDataset, ServiceNowObjectDataset, SharePointOnlineListResourceDataset, ShopifyObjectDataset, SnowflakeDataset, SparkObjectDataset, SqlServerTableDataset, SquareObjectDataset, SybaseTableDataset, TeradataTableDataset, VerticaTableDataset, WebTableDataset, XeroObjectDataset, XmlDataset, ZohoObjectDataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - } - - _subtype_map = { - 'type': {'AmazonMWSObject': 'AmazonMWSObjectDataset', 'AmazonRedshiftTable': 'AmazonRedshiftTableDataset', 'AmazonS3Object': 'AmazonS3Dataset', 'Avro': 'AvroDataset', 'AzureBlob': 'AzureBlobDataset', 'AzureBlobFSFile': 'AzureBlobFSDataset', 'AzureDataExplorerTable': 'AzureDataExplorerTableDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'AzureDatabricksDeltaLakeDataset': 'AzureDatabricksDeltaLakeDataset', 'AzureMariaDBTable': 'AzureMariaDBTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlMITable': 'AzureSqlMITableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'Binary': 'BinaryDataset', 'CassandraTable': 'CassandraTableDataset', 'CommonDataServiceForAppsEntity': 'CommonDataServiceForAppsEntityDataset', 'ConcurObject': 'ConcurObjectDataset', 'CosmosDbMongoDbApiCollection': 'CosmosDbMongoDbApiCollectionDataset', 'CosmosDbSqlApiCollection': 'CosmosDbSqlApiCollectionDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'CustomDataset': 'CustomDataset', 'Db2Table': 'Db2TableDataset', 'DelimitedText': 'DelimitedTextDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'DrillTable': 'DrillTableDataset', 'DynamicsAXResource': 'DynamicsAXResourceDataset', 'DynamicsCrmEntity': 'DynamicsCrmEntityDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'EloquaObject': 'EloquaObjectDataset', 'Excel': 'ExcelDataset', 'FileShare': 'FileShareDataset', 'GoogleAdWordsObject': 'GoogleAdWordsObjectDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'HBaseObject': 'HBaseObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HttpFile': 'HttpDataset', 'HubspotObject': 'HubspotObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'InformixTable': 'InformixTableDataset', 'JiraObject': 'JiraObjectDataset', 'Json': 'JsonDataset', 'MagentoObject': 'MagentoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MarketoObject': 'MarketoObjectDataset', 'MicrosoftAccessTable': 'MicrosoftAccessTableDataset', 'MongoDbAtlasCollection': 'MongoDbAtlasCollectionDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'MongoDbV2Collection': 'MongoDbV2CollectionDataset', 'MySqlTable': 'MySqlTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ODataResource': 'ODataResourceDataset', 'OdbcTable': 'OdbcTableDataset', 'Office365Table': 'Office365Dataset', 'OracleServiceCloudObject': 'OracleServiceCloudObjectDataset', 'OracleTable': 'OracleTableDataset', 'Orc': 'OrcDataset', 'Parquet': 'ParquetDataset', 'PaypalObject': 'PaypalObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PostgreSqlTable': 'PostgreSqlTableDataset', 'PrestoObject': 'PrestoObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'ResponsysObject': 'ResponsysObjectDataset', 'RestResource': 'RestResourceDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'SalesforceServiceCloudObject': 'SalesforceServiceCloudObjectDataset', 'SapBwCube': 'SapBwCubeDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapHanaTable': 'SapHanaTableDataset', 'SapOpenHubTable': 'SapOpenHubTableDataset', 'SapTableResource': 'SapTableResourceDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'SharePointOnlineListResource': 'SharePointOnlineListResourceDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'SnowflakeTable': 'SnowflakeDataset', 'SparkObject': 'SparkObjectDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SquareObject': 'SquareObjectDataset', 'SybaseTable': 'SybaseTableDataset', 'TeradataTable': 'TeradataTableDataset', 'VerticaTable': 'VerticaTableDataset', 'WebTable': 'WebTableDataset', 'XeroObject': 'XeroObjectDataset', 'Xml': 'XmlDataset', 'ZohoObject': 'ZohoObjectDataset'} - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - **kwargs - ): - super(Dataset, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = 'Dataset' # type: str - self.description = description - self.structure = structure - self.schema = schema - self.linked_service_name = linked_service_name - self.parameters = parameters - self.annotations = annotations - self.folder = folder - - -class AmazonMWSObjectDataset(Dataset): - """Amazon Marketplace Web Service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(AmazonMWSObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'AmazonMWSObject' # type: str - self.table_name = table_name - - -class CopySource(msrest.serialization.Model): - """A copy activity source. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AvroSource, AzureBlobFSSource, AzureDataExplorerSource, AzureDataLakeStoreSource, AzureDatabricksDeltaLakeSource, BinarySource, BlobSource, CommonDataServiceForAppsSource, CosmosDbMongoDbApiSource, CosmosDbSqlApiSource, DelimitedTextSource, DocumentDbCollectionSource, DynamicsCrmSource, DynamicsSource, ExcelSource, FileSystemSource, HdfsSource, HttpSource, JsonSource, MicrosoftAccessSource, MongoDbAtlasSource, MongoDbSource, MongoDbV2Source, ODataSource, Office365Source, OracleSource, OrcSource, ParquetSource, RelationalSource, RestSource, SalesforceServiceCloudSource, SharePointOnlineListSource, SnowflakeSource, TabularSource, WebSource, XmlSource. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - } - - _subtype_map = { - 'type': {'AvroSource': 'AvroSource', 'AzureBlobFSSource': 'AzureBlobFSSource', 'AzureDataExplorerSource': 'AzureDataExplorerSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'AzureDatabricksDeltaLakeSource': 'AzureDatabricksDeltaLakeSource', 'BinarySource': 'BinarySource', 'BlobSource': 'BlobSource', 'CommonDataServiceForAppsSource': 'CommonDataServiceForAppsSource', 'CosmosDbMongoDbApiSource': 'CosmosDbMongoDbApiSource', 'CosmosDbSqlApiSource': 'CosmosDbSqlApiSource', 'DelimitedTextSource': 'DelimitedTextSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'DynamicsCrmSource': 'DynamicsCrmSource', 'DynamicsSource': 'DynamicsSource', 'ExcelSource': 'ExcelSource', 'FileSystemSource': 'FileSystemSource', 'HdfsSource': 'HdfsSource', 'HttpSource': 'HttpSource', 'JsonSource': 'JsonSource', 'MicrosoftAccessSource': 'MicrosoftAccessSource', 'MongoDbAtlasSource': 'MongoDbAtlasSource', 'MongoDbSource': 'MongoDbSource', 'MongoDbV2Source': 'MongoDbV2Source', 'ODataSource': 'ODataSource', 'Office365Source': 'Office365Source', 'OracleSource': 'OracleSource', 'OrcSource': 'OrcSource', 'ParquetSource': 'ParquetSource', 'RelationalSource': 'RelationalSource', 'RestSource': 'RestSource', 'SalesforceServiceCloudSource': 'SalesforceServiceCloudSource', 'SharePointOnlineListSource': 'SharePointOnlineListSource', 'SnowflakeSource': 'SnowflakeSource', 'TabularSource': 'TabularSource', 'WebSource': 'WebSource', 'XmlSource': 'XmlSource'} - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - **kwargs - ): - super(CopySource, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = 'CopySource' # type: str - self.source_retry_count = source_retry_count - self.source_retry_wait = source_retry_wait - self.max_concurrent_connections = max_concurrent_connections - - -class TabularSource(CopySource): - """Copy activity sources of tabular type. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonMWSSource, AmazonRedshiftSource, AzureMariaDBSource, AzureMySqlSource, AzurePostgreSqlSource, AzureSqlSource, AzureTableSource, CassandraSource, ConcurSource, CouchbaseSource, Db2Source, DrillSource, DynamicsAXSource, EloquaSource, GoogleAdWordsSource, GoogleBigQuerySource, GreenplumSource, HBaseSource, HiveSource, HubspotSource, ImpalaSource, InformixSource, JiraSource, MagentoSource, MariaDBSource, MarketoSource, MySqlSource, NetezzaSource, OdbcSource, OracleServiceCloudSource, PaypalSource, PhoenixSource, PostgreSqlSource, PrestoSource, QuickBooksSource, ResponsysSource, SalesforceMarketingCloudSource, SalesforceSource, SapBwSource, SapCloudForCustomerSource, SapEccSource, SapHanaSource, SapOpenHubSource, SapTableSource, ServiceNowSource, ShopifySource, SparkSource, SqlDWSource, SqlMISource, SqlServerSource, SqlSource, SquareSource, SybaseSource, TeradataSource, VerticaSource, XeroSource, ZohoSource. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - _subtype_map = { - 'type': {'AmazonMWSSource': 'AmazonMWSSource', 'AmazonRedshiftSource': 'AmazonRedshiftSource', 'AzureMariaDBSource': 'AzureMariaDBSource', 'AzureMySqlSource': 'AzureMySqlSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AzureSqlSource': 'AzureSqlSource', 'AzureTableSource': 'AzureTableSource', 'CassandraSource': 'CassandraSource', 'ConcurSource': 'ConcurSource', 'CouchbaseSource': 'CouchbaseSource', 'Db2Source': 'Db2Source', 'DrillSource': 'DrillSource', 'DynamicsAXSource': 'DynamicsAXSource', 'EloquaSource': 'EloquaSource', 'GoogleAdWordsSource': 'GoogleAdWordsSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'GreenplumSource': 'GreenplumSource', 'HBaseSource': 'HBaseSource', 'HiveSource': 'HiveSource', 'HubspotSource': 'HubspotSource', 'ImpalaSource': 'ImpalaSource', 'InformixSource': 'InformixSource', 'JiraSource': 'JiraSource', 'MagentoSource': 'MagentoSource', 'MariaDBSource': 'MariaDBSource', 'MarketoSource': 'MarketoSource', 'MySqlSource': 'MySqlSource', 'NetezzaSource': 'NetezzaSource', 'OdbcSource': 'OdbcSource', 'OracleServiceCloudSource': 'OracleServiceCloudSource', 'PaypalSource': 'PaypalSource', 'PhoenixSource': 'PhoenixSource', 'PostgreSqlSource': 'PostgreSqlSource', 'PrestoSource': 'PrestoSource', 'QuickBooksSource': 'QuickBooksSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'SalesforceSource': 'SalesforceSource', 'SapBwSource': 'SapBwSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SapEccSource': 'SapEccSource', 'SapHanaSource': 'SapHanaSource', 'SapOpenHubSource': 'SapOpenHubSource', 'SapTableSource': 'SapTableSource', 'ServiceNowSource': 'ServiceNowSource', 'ShopifySource': 'ShopifySource', 'SparkSource': 'SparkSource', 'SqlDWSource': 'SqlDWSource', 'SqlMISource': 'SqlMISource', 'SqlServerSource': 'SqlServerSource', 'SqlSource': 'SqlSource', 'SquareSource': 'SquareSource', 'SybaseSource': 'SybaseSource', 'TeradataSource': 'TeradataSource', 'VerticaSource': 'VerticaSource', 'XeroSource': 'XeroSource', 'ZohoSource': 'ZohoSource'} - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(TabularSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'TabularSource' # type: str - self.query_timeout = query_timeout - self.additional_columns = additional_columns - - -class AmazonMWSSource(TabularSource): - """A copy activity Amazon Marketplace Web Service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(AmazonMWSSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'AmazonMWSSource' # type: str - self.query = query - - -class AmazonRedshiftLinkedService(LinkedService): - """Linked service for Amazon Redshift. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param server: Required. The name of the Amazon Redshift server. Type: string (or Expression - with resultType string). - :type server: any - :param username: The username of the Amazon Redshift source. Type: string (or Expression with - resultType string). - :type username: any - :param password: The password of the Amazon Redshift source. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param database: Required. The database name of the Amazon Redshift source. Type: string (or - Expression with resultType string). - :type database: any - :param port: The TCP port number that the Amazon Redshift server uses to listen for client - connections. The default value is 5439. Type: integer (or Expression with resultType integer). - :type port: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'server': {'required': True}, - 'database': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - server: Any, - database: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - port: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AmazonRedshiftLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AmazonRedshift' # type: str - self.server = server - self.username = username - self.password = password - self.database = database - self.port = port - self.encrypted_credential = encrypted_credential - - -class AmazonRedshiftSource(TabularSource): - """A copy activity source for Amazon Redshift Source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - :param redshift_unload_settings: The Amazon S3 settings needed for the interim Amazon S3 when - copying from Amazon Redshift with unload. With this, data from Amazon Redshift source will be - unloaded into S3 first and then copied into the targeted sink from the interim S3. - :type redshift_unload_settings: - ~azure.synapse.artifacts.v2020_12_01.models.RedshiftUnloadSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'redshift_unload_settings': {'key': 'redshiftUnloadSettings', 'type': 'RedshiftUnloadSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - redshift_unload_settings: Optional["RedshiftUnloadSettings"] = None, - **kwargs - ): - super(AmazonRedshiftSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'AmazonRedshiftSource' # type: str - self.query = query - self.redshift_unload_settings = redshift_unload_settings - - -class AmazonRedshiftTableDataset(Dataset): - """The Amazon Redshift table dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The Amazon Redshift table name. Type: string (or Expression with resultType - string). - :type table: any - :param schema_type_properties_schema: The Amazon Redshift schema name. Type: string (or - Expression with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - table: Optional[Any] = None, - schema_type_properties_schema: Optional[Any] = None, - **kwargs - ): - super(AmazonRedshiftTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'AmazonRedshiftTable' # type: str - self.table_name = table_name - self.table = table - self.schema_type_properties_schema = schema_type_properties_schema - - -class AmazonS3Dataset(Dataset): - """A single Amazon Simple Storage Service (S3) object or a set of S3 objects. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param bucket_name: Required. The name of the Amazon S3 bucket. Type: string (or Expression - with resultType string). - :type bucket_name: any - :param key: The key of the Amazon S3 object. Type: string (or Expression with resultType - string). - :type key: any - :param prefix: The prefix filter for the S3 object name. Type: string (or Expression with - resultType string). - :type prefix: any - :param version: The version for the S3 object. Type: string (or Expression with resultType - string). - :type version: any - :param modified_datetime_start: The start of S3 object's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of S3 object's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_end: any - :param format: The format of files. - :type format: ~azure.synapse.artifacts.v2020_12_01.models.DatasetStorageFormat - :param compression: The data compression method used for the Amazon S3 object. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'bucket_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'bucket_name': {'key': 'typeProperties.bucketName', 'type': 'object'}, - 'key': {'key': 'typeProperties.key', 'type': 'object'}, - 'prefix': {'key': 'typeProperties.prefix', 'type': 'object'}, - 'version': {'key': 'typeProperties.version', 'type': 'object'}, - 'modified_datetime_start': {'key': 'typeProperties.modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'typeProperties.modifiedDatetimeEnd', 'type': 'object'}, - 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - bucket_name: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - key: Optional[Any] = None, - prefix: Optional[Any] = None, - version: Optional[Any] = None, - modified_datetime_start: Optional[Any] = None, - modified_datetime_end: Optional[Any] = None, - format: Optional["DatasetStorageFormat"] = None, - compression: Optional["DatasetCompression"] = None, - **kwargs - ): - super(AmazonS3Dataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'AmazonS3Object' # type: str - self.bucket_name = bucket_name - self.key = key - self.prefix = prefix - self.version = version - self.modified_datetime_start = modified_datetime_start - self.modified_datetime_end = modified_datetime_end - self.format = format - self.compression = compression - - -class AmazonS3LinkedService(LinkedService): - """Linked service for Amazon S3. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param authentication_type: The authentication type of S3. Allowed value: AccessKey (default) - or TemporarySecurityCredentials. Type: string (or Expression with resultType string). - :type authentication_type: any - :param access_key_id: The access key identifier of the Amazon S3 Identity and Access Management - (IAM) user. Type: string (or Expression with resultType string). - :type access_key_id: any - :param secret_access_key: The secret access key of the Amazon S3 Identity and Access Management - (IAM) user. - :type secret_access_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param service_url: This value specifies the endpoint to access with the S3 Connector. This is - an optional property; change it only if you want to try a different service endpoint or want to - switch between https and http. Type: string (or Expression with resultType string). - :type service_url: any - :param session_token: The session token for the S3 temporary security credential. - :type session_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, - 'access_key_id': {'key': 'typeProperties.accessKeyId', 'type': 'object'}, - 'secret_access_key': {'key': 'typeProperties.secretAccessKey', 'type': 'SecretBase'}, - 'service_url': {'key': 'typeProperties.serviceUrl', 'type': 'object'}, - 'session_token': {'key': 'typeProperties.sessionToken', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - authentication_type: Optional[Any] = None, - access_key_id: Optional[Any] = None, - secret_access_key: Optional["SecretBase"] = None, - service_url: Optional[Any] = None, - session_token: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AmazonS3LinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AmazonS3' # type: str - self.authentication_type = authentication_type - self.access_key_id = access_key_id - self.secret_access_key = secret_access_key - self.service_url = service_url - self.session_token = session_token - self.encrypted_credential = encrypted_credential - - -class DatasetLocation(msrest.serialization.Model): - """Dataset location. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonS3Location, AzureBlobFSLocation, AzureBlobStorageLocation, AzureDataLakeStoreLocation, AzureFileStorageLocation, FileServerLocation, FtpServerLocation, GoogleCloudStorageLocation, HdfsLocation, HttpServerLocation, SftpLocation. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - } - - _subtype_map = { - 'type': {'AmazonS3Location': 'AmazonS3Location', 'AzureBlobFSLocation': 'AzureBlobFSLocation', 'AzureBlobStorageLocation': 'AzureBlobStorageLocation', 'AzureDataLakeStoreLocation': 'AzureDataLakeStoreLocation', 'AzureFileStorageLocation': 'AzureFileStorageLocation', 'FileServerLocation': 'FileServerLocation', 'FtpServerLocation': 'FtpServerLocation', 'GoogleCloudStorageLocation': 'GoogleCloudStorageLocation', 'HdfsLocation': 'HdfsLocation', 'HttpServerLocation': 'HttpServerLocation', 'SftpLocation': 'SftpLocation'} - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - folder_path: Optional[Any] = None, - file_name: Optional[Any] = None, - **kwargs - ): - super(DatasetLocation, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = 'DatasetLocation' # type: str - self.folder_path = folder_path - self.file_name = file_name - - -class AmazonS3Location(DatasetLocation): - """The location of amazon S3 dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - :param bucket_name: Specify the bucketName of amazon S3. Type: string (or Expression with - resultType string). - :type bucket_name: any - :param version: Specify the version of amazon S3. Type: string (or Expression with resultType - string). - :type version: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - 'bucket_name': {'key': 'bucketName', 'type': 'object'}, - 'version': {'key': 'version', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - folder_path: Optional[Any] = None, - file_name: Optional[Any] = None, - bucket_name: Optional[Any] = None, - version: Optional[Any] = None, - **kwargs - ): - super(AmazonS3Location, self).__init__(additional_properties=additional_properties, folder_path=folder_path, file_name=file_name, **kwargs) - self.type = 'AmazonS3Location' # type: str - self.bucket_name = bucket_name - self.version = version - - -class StoreReadSettings(msrest.serialization.Model): - """Connector read setting. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonS3ReadSettings, AzureBlobFSReadSettings, AzureBlobStorageReadSettings, AzureDataLakeStoreReadSettings, AzureFileStorageReadSettings, FileServerReadSettings, FtpReadSettings, GoogleCloudStorageReadSettings, HdfsReadSettings, HttpReadSettings, SftpReadSettings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - } - - _subtype_map = { - 'type': {'AmazonS3ReadSettings': 'AmazonS3ReadSettings', 'AzureBlobFSReadSettings': 'AzureBlobFSReadSettings', 'AzureBlobStorageReadSettings': 'AzureBlobStorageReadSettings', 'AzureDataLakeStoreReadSettings': 'AzureDataLakeStoreReadSettings', 'AzureFileStorageReadSettings': 'AzureFileStorageReadSettings', 'FileServerReadSettings': 'FileServerReadSettings', 'FtpReadSettings': 'FtpReadSettings', 'GoogleCloudStorageReadSettings': 'GoogleCloudStorageReadSettings', 'HdfsReadSettings': 'HdfsReadSettings', 'HttpReadSettings': 'HttpReadSettings', 'SftpReadSettings': 'SftpReadSettings'} - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - **kwargs - ): - super(StoreReadSettings, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = 'StoreReadSettings' # type: str - self.max_concurrent_connections = max_concurrent_connections - - -class AmazonS3ReadSettings(StoreReadSettings): - """Azure data lake store read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: AmazonS3 wildcardFolderPath. Type: string (or Expression with - resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: AmazonS3 wildcardFileName. Type: string (or Expression with - resultType string). - :type wildcard_file_name: any - :param prefix: The prefix filter for the S3 object name. Type: string (or Expression with - resultType string). - :type prefix: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'prefix': {'key': 'prefix', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - 'modified_datetime_start': {'key': 'modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'modifiedDatetimeEnd', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - recursive: Optional[Any] = None, - wildcard_folder_path: Optional[Any] = None, - wildcard_file_name: Optional[Any] = None, - prefix: Optional[Any] = None, - file_list_path: Optional[Any] = None, - enable_partition_discovery: Optional[bool] = None, - partition_root_path: Optional[Any] = None, - delete_files_after_completion: Optional[Any] = None, - modified_datetime_start: Optional[Any] = None, - modified_datetime_end: Optional[Any] = None, - **kwargs - ): - super(AmazonS3ReadSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AmazonS3ReadSettings' # type: str - self.recursive = recursive - self.wildcard_folder_path = wildcard_folder_path - self.wildcard_file_name = wildcard_file_name - self.prefix = prefix - self.file_list_path = file_list_path - self.enable_partition_discovery = enable_partition_discovery - self.partition_root_path = partition_root_path - self.delete_files_after_completion = delete_files_after_completion - self.modified_datetime_start = modified_datetime_start - self.modified_datetime_end = modified_datetime_end - - -class ControlActivity(Activity): - """Base class for all control activities like IfCondition, ForEach , Until. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AppendVariableActivity, ExecutePipelineActivity, FilterActivity, ForEachActivity, IfConditionActivity, SetVariableActivity, SwitchActivity, UntilActivity, ValidationActivity, WaitActivity, WebHookActivity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - } - - _subtype_map = { - 'type': {'AppendVariable': 'AppendVariableActivity', 'ExecutePipeline': 'ExecutePipelineActivity', 'Filter': 'FilterActivity', 'ForEach': 'ForEachActivity', 'IfCondition': 'IfConditionActivity', 'SetVariable': 'SetVariableActivity', 'Switch': 'SwitchActivity', 'Until': 'UntilActivity', 'Validation': 'ValidationActivity', 'Wait': 'WaitActivity', 'WebHook': 'WebHookActivity'} - } - - def __init__( - self, - *, - name: str, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - **kwargs - ): - super(ControlActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) - self.type = 'Container' # type: str - - -class AppendVariableActivity(ControlActivity): - """Append value for a Variable of type Array. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param variable_name: Name of the variable whose value needs to be appended to. - :type variable_name: str - :param value: Value to be appended. Could be a static value or Expression. - :type value: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'variable_name': {'key': 'typeProperties.variableName', 'type': 'str'}, - 'value': {'key': 'typeProperties.value', 'type': 'object'}, - } - - def __init__( - self, - *, - name: str, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - variable_name: Optional[str] = None, - value: Optional[Any] = None, - **kwargs - ): - super(AppendVariableActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) - self.type = 'AppendVariable' # type: str - self.variable_name = variable_name - self.value = value - - -class ArtifactRenameRequest(msrest.serialization.Model): - """Request body structure for rename artifact. - - :param new_name: New name of the artifact. - :type new_name: str - """ - - _validation = { - 'new_name': {'max_length': 260, 'min_length': 1, 'pattern': r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'}, - } - - _attribute_map = { - 'new_name': {'key': 'newName', 'type': 'str'}, - } - - def __init__( - self, - *, - new_name: Optional[str] = None, - **kwargs - ): - super(ArtifactRenameRequest, self).__init__(**kwargs) - self.new_name = new_name - - -class AutoPauseProperties(msrest.serialization.Model): - """Auto-pausing properties of a Big Data pool powered by Apache Spark. - - :param delay_in_minutes: Number of minutes of idle time before the Big Data pool is - automatically paused. - :type delay_in_minutes: int - :param enabled: Whether auto-pausing is enabled for the Big Data pool. - :type enabled: bool - """ - - _attribute_map = { - 'delay_in_minutes': {'key': 'delayInMinutes', 'type': 'int'}, - 'enabled': {'key': 'enabled', 'type': 'bool'}, - } - - def __init__( - self, - *, - delay_in_minutes: Optional[int] = None, - enabled: Optional[bool] = None, - **kwargs - ): - super(AutoPauseProperties, self).__init__(**kwargs) - self.delay_in_minutes = delay_in_minutes - self.enabled = enabled - - -class AutoScaleProperties(msrest.serialization.Model): - """Auto-scaling properties of a Big Data pool powered by Apache Spark. - - :param min_node_count: The minimum number of nodes the Big Data pool can support. - :type min_node_count: int - :param enabled: Whether automatic scaling is enabled for the Big Data pool. - :type enabled: bool - :param max_node_count: The maximum number of nodes the Big Data pool can support. - :type max_node_count: int - """ - - _attribute_map = { - 'min_node_count': {'key': 'minNodeCount', 'type': 'int'}, - 'enabled': {'key': 'enabled', 'type': 'bool'}, - 'max_node_count': {'key': 'maxNodeCount', 'type': 'int'}, - } - - def __init__( - self, - *, - min_node_count: Optional[int] = None, - enabled: Optional[bool] = None, - max_node_count: Optional[int] = None, - **kwargs - ): - super(AutoScaleProperties, self).__init__(**kwargs) - self.min_node_count = min_node_count - self.enabled = enabled - self.max_node_count = max_node_count - - -class AvroDataset(Dataset): - """Avro dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param location: The location of the avro storage. - :type location: ~azure.synapse.artifacts.v2020_12_01.models.DatasetLocation - :param avro_compression_codec: A string from AvroCompressionCodecEnum or an expression. - :type avro_compression_codec: any - :param avro_compression_level: - :type avro_compression_level: int - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'avro_compression_level': {'maximum': 9, 'minimum': 1}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'location': {'key': 'typeProperties.location', 'type': 'DatasetLocation'}, - 'avro_compression_codec': {'key': 'typeProperties.avroCompressionCodec', 'type': 'object'}, - 'avro_compression_level': {'key': 'typeProperties.avroCompressionLevel', 'type': 'int'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - location: Optional["DatasetLocation"] = None, - avro_compression_codec: Optional[Any] = None, - avro_compression_level: Optional[int] = None, - **kwargs - ): - super(AvroDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'Avro' # type: str - self.location = location - self.avro_compression_codec = avro_compression_codec - self.avro_compression_level = avro_compression_level - - -class DatasetStorageFormat(msrest.serialization.Model): - """The format definition of a storage. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AvroFormat, JsonFormat, OrcFormat, ParquetFormat, TextFormat. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage format.Constant filled by server. - :type type: str - :param serializer: Serializer. Type: string (or Expression with resultType string). - :type serializer: any - :param deserializer: Deserializer. Type: string (or Expression with resultType string). - :type deserializer: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'serializer': {'key': 'serializer', 'type': 'object'}, - 'deserializer': {'key': 'deserializer', 'type': 'object'}, - } - - _subtype_map = { - 'type': {'AvroFormat': 'AvroFormat', 'JsonFormat': 'JsonFormat', 'OrcFormat': 'OrcFormat', 'ParquetFormat': 'ParquetFormat', 'TextFormat': 'TextFormat'} - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - serializer: Optional[Any] = None, - deserializer: Optional[Any] = None, - **kwargs - ): - super(DatasetStorageFormat, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = 'DatasetStorageFormat' # type: str - self.serializer = serializer - self.deserializer = deserializer - - -class AvroFormat(DatasetStorageFormat): - """The data stored in Avro format. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage format.Constant filled by server. - :type type: str - :param serializer: Serializer. Type: string (or Expression with resultType string). - :type serializer: any - :param deserializer: Deserializer. Type: string (or Expression with resultType string). - :type deserializer: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'serializer': {'key': 'serializer', 'type': 'object'}, - 'deserializer': {'key': 'deserializer', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - serializer: Optional[Any] = None, - deserializer: Optional[Any] = None, - **kwargs - ): - super(AvroFormat, self).__init__(additional_properties=additional_properties, serializer=serializer, deserializer=deserializer, **kwargs) - self.type = 'AvroFormat' # type: str - - -class CopySink(msrest.serialization.Model): - """A copy activity sink. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AvroSink, AzureBlobFSSink, AzureDataExplorerSink, AzureDataLakeStoreSink, AzureDatabricksDeltaLakeSink, AzureMySqlSink, AzurePostgreSqlSink, AzureQueueSink, AzureSearchIndexSink, AzureSqlSink, AzureTableSink, BinarySink, BlobSink, CommonDataServiceForAppsSink, CosmosDbMongoDbApiSink, CosmosDbSqlApiSink, DelimitedTextSink, DocumentDbCollectionSink, DynamicsCrmSink, DynamicsSink, FileSystemSink, InformixSink, JsonSink, MicrosoftAccessSink, OdbcSink, OracleSink, OrcSink, ParquetSink, RestSink, SalesforceServiceCloudSink, SalesforceSink, SapCloudForCustomerSink, SnowflakeSink, SqlDWSink, SqlMISink, SqlServerSink, SqlSink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - } - - _subtype_map = { - 'type': {'AvroSink': 'AvroSink', 'AzureBlobFSSink': 'AzureBlobFSSink', 'AzureDataExplorerSink': 'AzureDataExplorerSink', 'AzureDataLakeStoreSink': 'AzureDataLakeStoreSink', 'AzureDatabricksDeltaLakeSink': 'AzureDatabricksDeltaLakeSink', 'AzureMySqlSink': 'AzureMySqlSink', 'AzurePostgreSqlSink': 'AzurePostgreSqlSink', 'AzureQueueSink': 'AzureQueueSink', 'AzureSearchIndexSink': 'AzureSearchIndexSink', 'AzureSqlSink': 'AzureSqlSink', 'AzureTableSink': 'AzureTableSink', 'BinarySink': 'BinarySink', 'BlobSink': 'BlobSink', 'CommonDataServiceForAppsSink': 'CommonDataServiceForAppsSink', 'CosmosDbMongoDbApiSink': 'CosmosDbMongoDbApiSink', 'CosmosDbSqlApiSink': 'CosmosDbSqlApiSink', 'DelimitedTextSink': 'DelimitedTextSink', 'DocumentDbCollectionSink': 'DocumentDbCollectionSink', 'DynamicsCrmSink': 'DynamicsCrmSink', 'DynamicsSink': 'DynamicsSink', 'FileSystemSink': 'FileSystemSink', 'InformixSink': 'InformixSink', 'JsonSink': 'JsonSink', 'MicrosoftAccessSink': 'MicrosoftAccessSink', 'OdbcSink': 'OdbcSink', 'OracleSink': 'OracleSink', 'OrcSink': 'OrcSink', 'ParquetSink': 'ParquetSink', 'RestSink': 'RestSink', 'SalesforceServiceCloudSink': 'SalesforceServiceCloudSink', 'SalesforceSink': 'SalesforceSink', 'SapCloudForCustomerSink': 'SapCloudForCustomerSink', 'SnowflakeSink': 'SnowflakeSink', 'SqlDWSink': 'SqlDWSink', 'SqlMISink': 'SqlMISink', 'SqlServerSink': 'SqlServerSink', 'SqlSink': 'SqlSink'} - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - **kwargs - ): - super(CopySink, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = 'CopySink' # type: str - self.write_batch_size = write_batch_size - self.write_batch_timeout = write_batch_timeout - self.sink_retry_count = sink_retry_count - self.sink_retry_wait = sink_retry_wait - self.max_concurrent_connections = max_concurrent_connections - - -class AvroSink(CopySink): - """A copy activity Avro sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Avro store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreWriteSettings - :param format_settings: Avro format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.AvroWriteSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreWriteSettings'}, - 'format_settings': {'key': 'formatSettings', 'type': 'AvroWriteSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - store_settings: Optional["StoreWriteSettings"] = None, - format_settings: Optional["AvroWriteSettings"] = None, - **kwargs - ): - super(AvroSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AvroSink' # type: str - self.store_settings = store_settings - self.format_settings = format_settings - - -class AvroSource(CopySource): - """A copy activity Avro source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Avro store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - store_settings: Optional["StoreReadSettings"] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(AvroSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AvroSource' # type: str - self.store_settings = store_settings - self.additional_columns = additional_columns - - -class FormatWriteSettings(msrest.serialization.Model): - """Format write settings. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AvroWriteSettings, DelimitedTextWriteSettings, JsonWriteSettings, OrcWriteSettings, ParquetWriteSettings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - _subtype_map = { - 'type': {'AvroWriteSettings': 'AvroWriteSettings', 'DelimitedTextWriteSettings': 'DelimitedTextWriteSettings', 'JsonWriteSettings': 'JsonWriteSettings', 'OrcWriteSettings': 'OrcWriteSettings', 'ParquetWriteSettings': 'ParquetWriteSettings'} - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(FormatWriteSettings, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = 'FormatWriteSettings' # type: str - - -class AvroWriteSettings(FormatWriteSettings): - """Avro write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param record_name: Top level record name in write result, which is required in AVRO spec. - :type record_name: str - :param record_namespace: Record namespace in the write result. - :type record_namespace: str - :param max_rows_per_file: Limit the written file's row count to be smaller than or equal to the - specified count. Type: integer (or Expression with resultType integer). - :type max_rows_per_file: any - :param file_name_prefix: Specifies the file name pattern - :code:``_:code:``.:code:`` when copy from non-file - based store without partitionOptions. Type: string (or Expression with resultType string). - :type file_name_prefix: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'record_name': {'key': 'recordName', 'type': 'str'}, - 'record_namespace': {'key': 'recordNamespace', 'type': 'str'}, - 'max_rows_per_file': {'key': 'maxRowsPerFile', 'type': 'object'}, - 'file_name_prefix': {'key': 'fileNamePrefix', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - record_name: Optional[str] = None, - record_namespace: Optional[str] = None, - max_rows_per_file: Optional[Any] = None, - file_name_prefix: Optional[Any] = None, - **kwargs - ): - super(AvroWriteSettings, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'AvroWriteSettings' # type: str - self.record_name = record_name - self.record_namespace = record_namespace - self.max_rows_per_file = max_rows_per_file - self.file_name_prefix = file_name_prefix - - -class AzureBatchLinkedService(LinkedService): - """Azure Batch linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param account_name: Required. The Azure Batch account name. Type: string (or Expression with - resultType string). - :type account_name: any - :param access_key: The Azure Batch account access key. - :type access_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param batch_uri: Required. The Azure Batch URI. Type: string (or Expression with resultType - string). - :type batch_uri: any - :param pool_name: Required. The Azure Batch pool name. Type: string (or Expression with - resultType string). - :type pool_name: any - :param linked_service_name: Required. The Azure Storage linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'account_name': {'required': True}, - 'batch_uri': {'required': True}, - 'pool_name': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'account_name': {'key': 'typeProperties.accountName', 'type': 'object'}, - 'access_key': {'key': 'typeProperties.accessKey', 'type': 'SecretBase'}, - 'batch_uri': {'key': 'typeProperties.batchUri', 'type': 'object'}, - 'pool_name': {'key': 'typeProperties.poolName', 'type': 'object'}, - 'linked_service_name': {'key': 'typeProperties.linkedServiceName', 'type': 'LinkedServiceReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - account_name: Any, - batch_uri: Any, - pool_name: Any, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - access_key: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AzureBatchLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureBatch' # type: str - self.account_name = account_name - self.access_key = access_key - self.batch_uri = batch_uri - self.pool_name = pool_name - self.linked_service_name = linked_service_name - self.encrypted_credential = encrypted_credential - - -class AzureBlobDataset(Dataset): - """The Azure Blob storage. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param folder_path: The path of the Azure Blob storage. Type: string (or Expression with - resultType string). - :type folder_path: any - :param table_root_location: The root of blob path. Type: string (or Expression with resultType - string). - :type table_root_location: any - :param file_name: The name of the Azure Blob. Type: string (or Expression with resultType - string). - :type file_name: any - :param modified_datetime_start: The start of Azure Blob's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of Azure Blob's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_end: any - :param format: The format of the Azure Blob storage. - :type format: ~azure.synapse.artifacts.v2020_12_01.models.DatasetStorageFormat - :param compression: The data compression method used for the blob storage. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, - 'table_root_location': {'key': 'typeProperties.tableRootLocation', 'type': 'object'}, - 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, - 'modified_datetime_start': {'key': 'typeProperties.modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'typeProperties.modifiedDatetimeEnd', 'type': 'object'}, - 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - folder_path: Optional[Any] = None, - table_root_location: Optional[Any] = None, - file_name: Optional[Any] = None, - modified_datetime_start: Optional[Any] = None, - modified_datetime_end: Optional[Any] = None, - format: Optional["DatasetStorageFormat"] = None, - compression: Optional["DatasetCompression"] = None, - **kwargs - ): - super(AzureBlobDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'AzureBlob' # type: str - self.folder_path = folder_path - self.table_root_location = table_root_location - self.file_name = file_name - self.modified_datetime_start = modified_datetime_start - self.modified_datetime_end = modified_datetime_end - self.format = format - self.compression = compression - - -class AzureBlobFSDataset(Dataset): - """The Azure Data Lake Storage Gen2 storage. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param folder_path: The path of the Azure Data Lake Storage Gen2 storage. Type: string (or - Expression with resultType string). - :type folder_path: any - :param file_name: The name of the Azure Data Lake Storage Gen2. Type: string (or Expression - with resultType string). - :type file_name: any - :param format: The format of the Azure Data Lake Storage Gen2 storage. - :type format: ~azure.synapse.artifacts.v2020_12_01.models.DatasetStorageFormat - :param compression: The data compression method used for the blob storage. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, - 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, - 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - folder_path: Optional[Any] = None, - file_name: Optional[Any] = None, - format: Optional["DatasetStorageFormat"] = None, - compression: Optional["DatasetCompression"] = None, - **kwargs - ): - super(AzureBlobFSDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'AzureBlobFSFile' # type: str - self.folder_path = folder_path - self.file_name = file_name - self.format = format - self.compression = compression - - -class AzureBlobFSLinkedService(LinkedService): - """Azure Data Lake Storage Gen2 linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param url: Required. Endpoint for the Azure Data Lake Storage Gen2 service. Type: string (or - Expression with resultType string). - :type url: any - :param account_key: Account key for the Azure Data Lake Storage Gen2 service. Type: string (or - Expression with resultType string). - :type account_key: any - :param service_principal_id: The ID of the application used to authenticate against the Azure - Data Lake Storage Gen2 account. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: The Key of the application used to authenticate against the Azure - Data Lake Storage Gen2 account. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: The name or ID of the tenant to which the service principal belongs. Type: - string (or Expression with resultType string). - :type tenant: any - :param azure_cloud_type: Indicates the azure cloud type of the service principle auth. Allowed - values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data - factory regions’ cloud type. Type: string (or Expression with resultType string). - :type azure_cloud_type: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'account_key': {'key': 'typeProperties.accountKey', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'azure_cloud_type': {'key': 'typeProperties.azureCloudType', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - url: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - account_key: Optional[Any] = None, - service_principal_id: Optional[Any] = None, - service_principal_key: Optional["SecretBase"] = None, - tenant: Optional[Any] = None, - azure_cloud_type: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AzureBlobFSLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureBlobFS' # type: str - self.url = url - self.account_key = account_key - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.azure_cloud_type = azure_cloud_type - self.encrypted_credential = encrypted_credential - - -class AzureBlobFSLocation(DatasetLocation): - """The location of azure blobFS dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - :param file_system: Specify the fileSystem of azure blobFS. Type: string (or Expression with - resultType string). - :type file_system: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - 'file_system': {'key': 'fileSystem', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - folder_path: Optional[Any] = None, - file_name: Optional[Any] = None, - file_system: Optional[Any] = None, - **kwargs - ): - super(AzureBlobFSLocation, self).__init__(additional_properties=additional_properties, folder_path=folder_path, file_name=file_name, **kwargs) - self.type = 'AzureBlobFSLocation' # type: str - self.file_system = file_system - - -class AzureBlobFSReadSettings(StoreReadSettings): - """Azure blobFS read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: Azure blobFS wildcardFolderPath. Type: string (or Expression with - resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: Azure blobFS wildcardFileName. Type: string (or Expression with - resultType string). - :type wildcard_file_name: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - 'modified_datetime_start': {'key': 'modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'modifiedDatetimeEnd', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - recursive: Optional[Any] = None, - wildcard_folder_path: Optional[Any] = None, - wildcard_file_name: Optional[Any] = None, - file_list_path: Optional[Any] = None, - enable_partition_discovery: Optional[bool] = None, - partition_root_path: Optional[Any] = None, - delete_files_after_completion: Optional[Any] = None, - modified_datetime_start: Optional[Any] = None, - modified_datetime_end: Optional[Any] = None, - **kwargs - ): - super(AzureBlobFSReadSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AzureBlobFSReadSettings' # type: str - self.recursive = recursive - self.wildcard_folder_path = wildcard_folder_path - self.wildcard_file_name = wildcard_file_name - self.file_list_path = file_list_path - self.enable_partition_discovery = enable_partition_discovery - self.partition_root_path = partition_root_path - self.delete_files_after_completion = delete_files_after_completion - self.modified_datetime_start = modified_datetime_start - self.modified_datetime_end = modified_datetime_end - - -class AzureBlobFSSink(CopySink): - """A copy activity Azure Data Lake Storage Gen2 sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - copy_behavior: Optional[Any] = None, - **kwargs - ): - super(AzureBlobFSSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AzureBlobFSSink' # type: str - self.copy_behavior = copy_behavior - - -class AzureBlobFSSource(CopySource): - """A copy activity Azure BlobFS source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param treat_empty_as_null: Treat empty as null. Type: boolean (or Expression with resultType - boolean). - :type treat_empty_as_null: any - :param skip_header_line_count: Number of header lines to skip from each blob. Type: integer (or - Expression with resultType integer). - :type skip_header_line_count: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'treat_empty_as_null': {'key': 'treatEmptyAsNull', 'type': 'object'}, - 'skip_header_line_count': {'key': 'skipHeaderLineCount', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - treat_empty_as_null: Optional[Any] = None, - skip_header_line_count: Optional[Any] = None, - recursive: Optional[Any] = None, - **kwargs - ): - super(AzureBlobFSSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AzureBlobFSSource' # type: str - self.treat_empty_as_null = treat_empty_as_null - self.skip_header_line_count = skip_header_line_count - self.recursive = recursive - - -class StoreWriteSettings(msrest.serialization.Model): - """Connector write settings. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AzureBlobFSWriteSettings, AzureBlobStorageWriteSettings, AzureDataLakeStoreWriteSettings, AzureFileStorageWriteSettings, FileServerWriteSettings, SftpWriteSettings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - } - - _subtype_map = { - 'type': {'AzureBlobFSWriteSettings': 'AzureBlobFSWriteSettings', 'AzureBlobStorageWriteSettings': 'AzureBlobStorageWriteSettings', 'AzureDataLakeStoreWriteSettings': 'AzureDataLakeStoreWriteSettings', 'AzureFileStorageWriteSettings': 'AzureFileStorageWriteSettings', 'FileServerWriteSettings': 'FileServerWriteSettings', 'SftpWriteSettings': 'SftpWriteSettings'} - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - copy_behavior: Optional[Any] = None, - **kwargs - ): - super(StoreWriteSettings, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = 'StoreWriteSettings' # type: str - self.max_concurrent_connections = max_concurrent_connections - self.copy_behavior = copy_behavior - - -class AzureBlobFSWriteSettings(StoreWriteSettings): - """Azure blobFS write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - :param block_size_in_mb: Indicates the block size(MB) when writing data to blob. Type: integer - (or Expression with resultType integer). - :type block_size_in_mb: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - 'block_size_in_mb': {'key': 'blockSizeInMB', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - copy_behavior: Optional[Any] = None, - block_size_in_mb: Optional[Any] = None, - **kwargs - ): - super(AzureBlobFSWriteSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) - self.type = 'AzureBlobFSWriteSettings' # type: str - self.block_size_in_mb = block_size_in_mb - - -class AzureBlobStorageLinkedService(LinkedService): - """The azure blob storage linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: The connection string. It is mutually exclusive with sasUri, - serviceEndpoint property. Type: string, SecureString or AzureKeyVaultSecretReference. - :type connection_string: any - :param account_key: The Azure key vault secret reference of accountKey in connection string. - :type account_key: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param sas_uri: SAS URI of the Azure Blob Storage resource. It is mutually exclusive with - connectionString, serviceEndpoint property. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type sas_uri: any - :param sas_token: The Azure key vault secret reference of sasToken in sas uri. - :type sas_token: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param service_endpoint: Blob service endpoint of the Azure Blob Storage resource. It is - mutually exclusive with connectionString, sasUri property. - :type service_endpoint: str - :param service_principal_id: The ID of the service principal used to authenticate against Azure - SQL Data Warehouse. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: The key of the service principal used to authenticate against - Azure SQL Data Warehouse. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: The name or ID of the tenant to which the service principal belongs. Type: - string (or Expression with resultType string). - :type tenant: any - :param azure_cloud_type: Indicates the azure cloud type of the service principle auth. Allowed - values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data - factory regions’ cloud type. Type: string (or Expression with resultType string). - :type azure_cloud_type: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'account_key': {'key': 'typeProperties.accountKey', 'type': 'AzureKeyVaultSecretReference'}, - 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'object'}, - 'sas_token': {'key': 'typeProperties.sasToken', 'type': 'AzureKeyVaultSecretReference'}, - 'service_endpoint': {'key': 'typeProperties.serviceEndpoint', 'type': 'str'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'azure_cloud_type': {'key': 'typeProperties.azureCloudType', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_string: Optional[Any] = None, - account_key: Optional["AzureKeyVaultSecretReference"] = None, - sas_uri: Optional[Any] = None, - sas_token: Optional["AzureKeyVaultSecretReference"] = None, - service_endpoint: Optional[str] = None, - service_principal_id: Optional[Any] = None, - service_principal_key: Optional["SecretBase"] = None, - tenant: Optional[Any] = None, - azure_cloud_type: Optional[Any] = None, - encrypted_credential: Optional[str] = None, - **kwargs - ): - super(AzureBlobStorageLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureBlobStorage' # type: str - self.connection_string = connection_string - self.account_key = account_key - self.sas_uri = sas_uri - self.sas_token = sas_token - self.service_endpoint = service_endpoint - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.azure_cloud_type = azure_cloud_type - self.encrypted_credential = encrypted_credential - - -class AzureBlobStorageLocation(DatasetLocation): - """The location of azure blob dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - :param container: Specify the container of azure blob. Type: string (or Expression with - resultType string). - :type container: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - 'container': {'key': 'container', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - folder_path: Optional[Any] = None, - file_name: Optional[Any] = None, - container: Optional[Any] = None, - **kwargs - ): - super(AzureBlobStorageLocation, self).__init__(additional_properties=additional_properties, folder_path=folder_path, file_name=file_name, **kwargs) - self.type = 'AzureBlobStorageLocation' # type: str - self.container = container - - -class AzureBlobStorageReadSettings(StoreReadSettings): - """Azure blob read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: Azure blob wildcardFolderPath. Type: string (or Expression with - resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: Azure blob wildcardFileName. Type: string (or Expression with - resultType string). - :type wildcard_file_name: any - :param prefix: The prefix filter for the Azure Blob name. Type: string (or Expression with - resultType string). - :type prefix: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'prefix': {'key': 'prefix', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - 'modified_datetime_start': {'key': 'modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'modifiedDatetimeEnd', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - recursive: Optional[Any] = None, - wildcard_folder_path: Optional[Any] = None, - wildcard_file_name: Optional[Any] = None, - prefix: Optional[Any] = None, - file_list_path: Optional[Any] = None, - enable_partition_discovery: Optional[bool] = None, - partition_root_path: Optional[Any] = None, - delete_files_after_completion: Optional[Any] = None, - modified_datetime_start: Optional[Any] = None, - modified_datetime_end: Optional[Any] = None, - **kwargs - ): - super(AzureBlobStorageReadSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AzureBlobStorageReadSettings' # type: str - self.recursive = recursive - self.wildcard_folder_path = wildcard_folder_path - self.wildcard_file_name = wildcard_file_name - self.prefix = prefix - self.file_list_path = file_list_path - self.enable_partition_discovery = enable_partition_discovery - self.partition_root_path = partition_root_path - self.delete_files_after_completion = delete_files_after_completion - self.modified_datetime_start = modified_datetime_start - self.modified_datetime_end = modified_datetime_end - - -class AzureBlobStorageWriteSettings(StoreWriteSettings): - """Azure blob write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - :param block_size_in_mb: Indicates the block size(MB) when writing data to blob. Type: integer - (or Expression with resultType integer). - :type block_size_in_mb: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - 'block_size_in_mb': {'key': 'blockSizeInMB', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - copy_behavior: Optional[Any] = None, - block_size_in_mb: Optional[Any] = None, - **kwargs - ): - super(AzureBlobStorageWriteSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) - self.type = 'AzureBlobStorageWriteSettings' # type: str - self.block_size_in_mb = block_size_in_mb - - -class AzureDatabricksDeltaLakeDataset(Dataset): - """Azure Databricks Delta Lake dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table: The name of delta table. Type: string (or Expression with resultType string). - :type table: any - :param database: The database name of delta table. Type: string (or Expression with resultType - string). - :type database: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table: Optional[Any] = None, - database: Optional[Any] = None, - **kwargs - ): - super(AzureDatabricksDeltaLakeDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'AzureDatabricksDeltaLakeDataset' # type: str - self.table = table - self.database = database - - -class ExportSettings(msrest.serialization.Model): - """Export command settings. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AzureDatabricksDeltaLakeExportCommand, SnowflakeExportCopyCommand. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The export setting type.Constant filled by server. - :type type: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - _subtype_map = { - 'type': {'AzureDatabricksDeltaLakeExportCommand': 'AzureDatabricksDeltaLakeExportCommand', 'SnowflakeExportCopyCommand': 'SnowflakeExportCopyCommand'} - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(ExportSettings, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = 'ExportSettings' # type: str - - -class AzureDatabricksDeltaLakeExportCommand(ExportSettings): - """Azure Databricks Delta Lake export command settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The export setting type.Constant filled by server. - :type type: str - :param date_format: Specify the date format for the csv in Azure Databricks Delta Lake Copy. - Type: string (or Expression with resultType string). - :type date_format: any - :param timestamp_format: Specify the timestamp format for the csv in Azure Databricks Delta - Lake Copy. Type: string (or Expression with resultType string). - :type timestamp_format: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'date_format': {'key': 'dateFormat', 'type': 'object'}, - 'timestamp_format': {'key': 'timestampFormat', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - date_format: Optional[Any] = None, - timestamp_format: Optional[Any] = None, - **kwargs - ): - super(AzureDatabricksDeltaLakeExportCommand, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'AzureDatabricksDeltaLakeExportCommand' # type: str - self.date_format = date_format - self.timestamp_format = timestamp_format - - -class ImportSettings(msrest.serialization.Model): - """Import command settings. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AzureDatabricksDeltaLakeImportCommand, SnowflakeImportCopyCommand. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The import setting type.Constant filled by server. - :type type: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - _subtype_map = { - 'type': {'AzureDatabricksDeltaLakeImportCommand': 'AzureDatabricksDeltaLakeImportCommand', 'SnowflakeImportCopyCommand': 'SnowflakeImportCopyCommand'} - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(ImportSettings, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = 'ImportSettings' # type: str - - -class AzureDatabricksDeltaLakeImportCommand(ImportSettings): - """Azure Databricks Delta Lake import command settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The import setting type.Constant filled by server. - :type type: str - :param date_format: Specify the date format for csv in Azure Databricks Delta Lake Copy. Type: - string (or Expression with resultType string). - :type date_format: any - :param timestamp_format: Specify the timestamp format for csv in Azure Databricks Delta Lake - Copy. Type: string (or Expression with resultType string). - :type timestamp_format: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'date_format': {'key': 'dateFormat', 'type': 'object'}, - 'timestamp_format': {'key': 'timestampFormat', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - date_format: Optional[Any] = None, - timestamp_format: Optional[Any] = None, - **kwargs - ): - super(AzureDatabricksDeltaLakeImportCommand, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'AzureDatabricksDeltaLakeImportCommand' # type: str - self.date_format = date_format - self.timestamp_format = timestamp_format - - -class AzureDatabricksDeltaLakeLinkedService(LinkedService): - """Azure Databricks Delta Lake linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param domain: Required. :code:``.azuredatabricks.net, domain name of your Databricks - deployment. Type: string (or Expression with resultType string). - :type domain: any - :param access_token: Required. Access token for databricks REST API. Refer to - https://docs.azuredatabricks.net/api/latest/authentication.html. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type access_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param cluster_id: The id of an existing interactive cluster that will be used for all runs of - this job. Type: string (or Expression with resultType string). - :type cluster_id: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'domain': {'required': True}, - 'access_token': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'domain': {'key': 'typeProperties.domain', 'type': 'object'}, - 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, - 'cluster_id': {'key': 'typeProperties.clusterId', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - domain: Any, - access_token: "SecretBase", - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - cluster_id: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AzureDatabricksDeltaLakeLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureDatabricksDeltaLake' # type: str - self.domain = domain - self.access_token = access_token - self.cluster_id = cluster_id - self.encrypted_credential = encrypted_credential - - -class AzureDatabricksDeltaLakeSink(CopySink): - """A copy activity Azure Databricks Delta Lake sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType - string). - :type pre_copy_script: any - :param import_settings: Azure Databricks Delta Lake import settings. - :type import_settings: - ~azure.synapse.artifacts.v2020_12_01.models.AzureDatabricksDeltaLakeImportCommand - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - 'import_settings': {'key': 'importSettings', 'type': 'AzureDatabricksDeltaLakeImportCommand'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - pre_copy_script: Optional[Any] = None, - import_settings: Optional["AzureDatabricksDeltaLakeImportCommand"] = None, - **kwargs - ): - super(AzureDatabricksDeltaLakeSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AzureDatabricksDeltaLakeSink' # type: str - self.pre_copy_script = pre_copy_script - self.import_settings = import_settings - - -class AzureDatabricksDeltaLakeSource(CopySource): - """A copy activity Azure Databricks Delta Lake source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: Azure Databricks Delta Lake Sql query. Type: string (or Expression with - resultType string). - :type query: any - :param export_settings: Azure Databricks Delta Lake export settings. - :type export_settings: - ~azure.synapse.artifacts.v2020_12_01.models.AzureDatabricksDeltaLakeExportCommand - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'export_settings': {'key': 'exportSettings', 'type': 'AzureDatabricksDeltaLakeExportCommand'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query: Optional[Any] = None, - export_settings: Optional["AzureDatabricksDeltaLakeExportCommand"] = None, - **kwargs - ): - super(AzureDatabricksDeltaLakeSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AzureDatabricksDeltaLakeSource' # type: str - self.query = query - self.export_settings = export_settings - - -class AzureDatabricksLinkedService(LinkedService): - """Azure Databricks linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param domain: Required. :code:``.azuredatabricks.net, domain name of your Databricks - deployment. Type: string (or Expression with resultType string). - :type domain: any - :param access_token: Access token for databricks REST API. Refer to - https://docs.azuredatabricks.net/api/latest/authentication.html. Type: string (or Expression - with resultType string). - :type access_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param authentication: Required to specify MSI, if using Workspace resource id for databricks - REST API. Type: string (or Expression with resultType string). - :type authentication: any - :param workspace_resource_id: Workspace resource id for databricks REST API. Type: string (or - Expression with resultType string). - :type workspace_resource_id: any - :param existing_cluster_id: The id of an existing interactive cluster that will be used for all - runs of this activity. Type: string (or Expression with resultType string). - :type existing_cluster_id: any - :param instance_pool_id: The id of an existing instance pool that will be used for all runs of - this activity. Type: string (or Expression with resultType string). - :type instance_pool_id: any - :param new_cluster_version: If not using an existing interactive cluster, this specifies the - Spark version of a new job cluster or instance pool nodes created for each run of this - activity. Required if instancePoolId is specified. Type: string (or Expression with resultType - string). - :type new_cluster_version: any - :param new_cluster_num_of_worker: If not using an existing interactive cluster, this specifies - the number of worker nodes to use for the new job cluster or instance pool. For new job - clusters, this a string-formatted Int32, like '1' means numOfWorker is 1 or '1:10' means - auto-scale from 1 (min) to 10 (max). For instance pools, this is a string-formatted Int32, and - can only specify a fixed number of worker nodes, such as '2'. Required if newClusterVersion is - specified. Type: string (or Expression with resultType string). - :type new_cluster_num_of_worker: any - :param new_cluster_node_type: The node type of the new job cluster. This property is required - if newClusterVersion is specified and instancePoolId is not specified. If instancePoolId is - specified, this property is ignored. Type: string (or Expression with resultType string). - :type new_cluster_node_type: any - :param new_cluster_spark_conf: A set of optional, user-specified Spark configuration key-value - pairs. - :type new_cluster_spark_conf: dict[str, any] - :param new_cluster_spark_env_vars: A set of optional, user-specified Spark environment - variables key-value pairs. - :type new_cluster_spark_env_vars: dict[str, any] - :param new_cluster_custom_tags: Additional tags for cluster resources. This property is ignored - in instance pool configurations. - :type new_cluster_custom_tags: dict[str, any] - :param new_cluster_log_destination: Specify a location to deliver Spark driver, worker, and - event logs. Type: string (or Expression with resultType string). - :type new_cluster_log_destination: any - :param new_cluster_driver_node_type: The driver node type for the new job cluster. This - property is ignored in instance pool configurations. Type: string (or Expression with - resultType string). - :type new_cluster_driver_node_type: any - :param new_cluster_init_scripts: User-defined initialization scripts for the new cluster. Type: - array of strings (or Expression with resultType array of strings). - :type new_cluster_init_scripts: any - :param new_cluster_enable_elastic_disk: Enable the elastic disk on the new cluster. This - property is now ignored, and takes the default elastic disk behavior in Databricks (elastic - disks are always enabled). Type: boolean (or Expression with resultType boolean). - :type new_cluster_enable_elastic_disk: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - :param policy_id: The policy id for limiting the ability to configure clusters based on a user - defined set of rules. Type: string (or Expression with resultType string). - :type policy_id: any - """ - - _validation = { - 'type': {'required': True}, - 'domain': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'domain': {'key': 'typeProperties.domain', 'type': 'object'}, - 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, - 'authentication': {'key': 'typeProperties.authentication', 'type': 'object'}, - 'workspace_resource_id': {'key': 'typeProperties.workspaceResourceId', 'type': 'object'}, - 'existing_cluster_id': {'key': 'typeProperties.existingClusterId', 'type': 'object'}, - 'instance_pool_id': {'key': 'typeProperties.instancePoolId', 'type': 'object'}, - 'new_cluster_version': {'key': 'typeProperties.newClusterVersion', 'type': 'object'}, - 'new_cluster_num_of_worker': {'key': 'typeProperties.newClusterNumOfWorker', 'type': 'object'}, - 'new_cluster_node_type': {'key': 'typeProperties.newClusterNodeType', 'type': 'object'}, - 'new_cluster_spark_conf': {'key': 'typeProperties.newClusterSparkConf', 'type': '{object}'}, - 'new_cluster_spark_env_vars': {'key': 'typeProperties.newClusterSparkEnvVars', 'type': '{object}'}, - 'new_cluster_custom_tags': {'key': 'typeProperties.newClusterCustomTags', 'type': '{object}'}, - 'new_cluster_log_destination': {'key': 'typeProperties.newClusterLogDestination', 'type': 'object'}, - 'new_cluster_driver_node_type': {'key': 'typeProperties.newClusterDriverNodeType', 'type': 'object'}, - 'new_cluster_init_scripts': {'key': 'typeProperties.newClusterInitScripts', 'type': 'object'}, - 'new_cluster_enable_elastic_disk': {'key': 'typeProperties.newClusterEnableElasticDisk', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - 'policy_id': {'key': 'typeProperties.policyId', 'type': 'object'}, - } - - def __init__( - self, - *, - domain: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - access_token: Optional["SecretBase"] = None, - authentication: Optional[Any] = None, - workspace_resource_id: Optional[Any] = None, - existing_cluster_id: Optional[Any] = None, - instance_pool_id: Optional[Any] = None, - new_cluster_version: Optional[Any] = None, - new_cluster_num_of_worker: Optional[Any] = None, - new_cluster_node_type: Optional[Any] = None, - new_cluster_spark_conf: Optional[Dict[str, Any]] = None, - new_cluster_spark_env_vars: Optional[Dict[str, Any]] = None, - new_cluster_custom_tags: Optional[Dict[str, Any]] = None, - new_cluster_log_destination: Optional[Any] = None, - new_cluster_driver_node_type: Optional[Any] = None, - new_cluster_init_scripts: Optional[Any] = None, - new_cluster_enable_elastic_disk: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - policy_id: Optional[Any] = None, - **kwargs - ): - super(AzureDatabricksLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureDatabricks' # type: str - self.domain = domain - self.access_token = access_token - self.authentication = authentication - self.workspace_resource_id = workspace_resource_id - self.existing_cluster_id = existing_cluster_id - self.instance_pool_id = instance_pool_id - self.new_cluster_version = new_cluster_version - self.new_cluster_num_of_worker = new_cluster_num_of_worker - self.new_cluster_node_type = new_cluster_node_type - self.new_cluster_spark_conf = new_cluster_spark_conf - self.new_cluster_spark_env_vars = new_cluster_spark_env_vars - self.new_cluster_custom_tags = new_cluster_custom_tags - self.new_cluster_log_destination = new_cluster_log_destination - self.new_cluster_driver_node_type = new_cluster_driver_node_type - self.new_cluster_init_scripts = new_cluster_init_scripts - self.new_cluster_enable_elastic_disk = new_cluster_enable_elastic_disk - self.encrypted_credential = encrypted_credential - self.policy_id = policy_id - - -class ExecutionActivity(Activity): - """Base class for all execution activities. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AzureDataExplorerCommandActivity, AzureFunctionActivity, AzureMLBatchExecutionActivity, AzureMLExecutePipelineActivity, AzureMLUpdateResourceActivity, CopyActivity, CustomActivity, DataLakeAnalyticsUSQLActivity, DatabricksNotebookActivity, DatabricksSparkJarActivity, DatabricksSparkPythonActivity, DeleteActivity, ExecuteDataFlowActivity, ExecuteSSISPackageActivity, GetMetadataActivity, HDInsightHiveActivity, HDInsightMapReduceActivity, HDInsightPigActivity, HDInsightSparkActivity, HDInsightStreamingActivity, LookupActivity, SynapseSparkJobDefinitionActivity, SqlServerStoredProcedureActivity, SynapseNotebookActivity, WebActivity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - } - - _subtype_map = { - 'type': {'AzureDataExplorerCommand': 'AzureDataExplorerCommandActivity', 'AzureFunctionActivity': 'AzureFunctionActivity', 'AzureMLBatchExecution': 'AzureMLBatchExecutionActivity', 'AzureMLExecutePipeline': 'AzureMLExecutePipelineActivity', 'AzureMLUpdateResource': 'AzureMLUpdateResourceActivity', 'Copy': 'CopyActivity', 'Custom': 'CustomActivity', 'DataLakeAnalyticsU-SQL': 'DataLakeAnalyticsUSQLActivity', 'DatabricksNotebook': 'DatabricksNotebookActivity', 'DatabricksSparkJar': 'DatabricksSparkJarActivity', 'DatabricksSparkPython': 'DatabricksSparkPythonActivity', 'Delete': 'DeleteActivity', 'ExecuteDataFlow': 'ExecuteDataFlowActivity', 'ExecuteSSISPackage': 'ExecuteSSISPackageActivity', 'GetMetadata': 'GetMetadataActivity', 'HDInsightHive': 'HDInsightHiveActivity', 'HDInsightMapReduce': 'HDInsightMapReduceActivity', 'HDInsightPig': 'HDInsightPigActivity', 'HDInsightSpark': 'HDInsightSparkActivity', 'HDInsightStreaming': 'HDInsightStreamingActivity', 'Lookup': 'LookupActivity', 'SparkJob': 'SynapseSparkJobDefinitionActivity', 'SqlServerStoredProcedure': 'SqlServerStoredProcedureActivity', 'SynapseNotebook': 'SynapseNotebookActivity', 'WebActivity': 'WebActivity'} - } - - def __init__( - self, - *, - name: str, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - **kwargs - ): - super(ExecutionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) - self.type = 'Execution' # type: str - self.linked_service_name = linked_service_name - self.policy = policy - - -class AzureDataExplorerCommandActivity(ExecutionActivity): - """Azure Data Explorer command activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param command: Required. A control command, according to the Azure Data Explorer command - syntax. Type: string (or Expression with resultType string). - :type command: any - :param command_timeout: Control command timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))..). - :type command_timeout: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'command': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'command': {'key': 'typeProperties.command', 'type': 'object'}, - 'command_timeout': {'key': 'typeProperties.commandTimeout', 'type': 'object'}, - } - - def __init__( - self, - *, - name: str, - command: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - command_timeout: Optional[Any] = None, - **kwargs - ): - super(AzureDataExplorerCommandActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'AzureDataExplorerCommand' # type: str - self.command = command - self.command_timeout = command_timeout - - -class AzureDataExplorerLinkedService(LinkedService): - """Azure Data Explorer (Kusto) linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param endpoint: Required. The endpoint of Azure Data Explorer (the engine's endpoint). URL - will be in the format https://:code:``.:code:``.kusto.windows.net. - Type: string (or Expression with resultType string). - :type endpoint: any - :param service_principal_id: Required. The ID of the service principal used to authenticate - against Azure Data Explorer. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: Required. The key of the service principal used to authenticate - against Kusto. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param database: Required. Database name for connection. Type: string (or Expression with - resultType string). - :type database: any - :param tenant: Required. The name or ID of the tenant to which the service principal belongs. - Type: string (or Expression with resultType string). - :type tenant: any - """ - - _validation = { - 'type': {'required': True}, - 'endpoint': {'required': True}, - 'service_principal_id': {'required': True}, - 'service_principal_key': {'required': True}, - 'database': {'required': True}, - 'tenant': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - } - - def __init__( - self, - *, - endpoint: Any, - service_principal_id: Any, - service_principal_key: "SecretBase", - database: Any, - tenant: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - **kwargs - ): - super(AzureDataExplorerLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureDataExplorer' # type: str - self.endpoint = endpoint - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.database = database - self.tenant = tenant - - -class AzureDataExplorerSink(CopySink): - """A copy activity Azure Data Explorer sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param ingestion_mapping_name: A name of a pre-created csv mapping that was defined on the - target Kusto table. Type: string. - :type ingestion_mapping_name: any - :param ingestion_mapping_as_json: An explicit column mapping description provided in a json - format. Type: string. - :type ingestion_mapping_as_json: any - :param flush_immediately: If set to true, any aggregation will be skipped. Default is false. - Type: boolean. - :type flush_immediately: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'ingestion_mapping_name': {'key': 'ingestionMappingName', 'type': 'object'}, - 'ingestion_mapping_as_json': {'key': 'ingestionMappingAsJson', 'type': 'object'}, - 'flush_immediately': {'key': 'flushImmediately', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - ingestion_mapping_name: Optional[Any] = None, - ingestion_mapping_as_json: Optional[Any] = None, - flush_immediately: Optional[Any] = None, - **kwargs - ): - super(AzureDataExplorerSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AzureDataExplorerSink' # type: str - self.ingestion_mapping_name = ingestion_mapping_name - self.ingestion_mapping_as_json = ingestion_mapping_as_json - self.flush_immediately = flush_immediately - - -class AzureDataExplorerSource(CopySource): - """A copy activity Azure Data Explorer (Kusto) source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: Required. Database query. Should be a Kusto Query Language (KQL) query. Type: - string (or Expression with resultType string). - :type query: any - :param no_truncation: The name of the Boolean option that controls whether truncation is - applied to result-sets that go beyond a certain row-count limit. - :type no_truncation: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])).. - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - 'query': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'no_truncation': {'key': 'noTruncation', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - query: Any, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - no_truncation: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(AzureDataExplorerSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AzureDataExplorerSource' # type: str - self.query = query - self.no_truncation = no_truncation - self.query_timeout = query_timeout - self.additional_columns = additional_columns - - -class AzureDataExplorerTableDataset(Dataset): - """The Azure Data Explorer (Kusto) dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table: The table name of the Azure Data Explorer database. Type: string (or Expression - with resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table: Optional[Any] = None, - **kwargs - ): - super(AzureDataExplorerTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'AzureDataExplorerTable' # type: str - self.table = table - - -class AzureDataLakeAnalyticsLinkedService(LinkedService): - """Azure Data Lake Analytics linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param account_name: Required. The Azure Data Lake Analytics account name. Type: string (or - Expression with resultType string). - :type account_name: any - :param service_principal_id: The ID of the application used to authenticate against the Azure - Data Lake Analytics account. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: The Key of the application used to authenticate against the Azure - Data Lake Analytics account. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: Required. The name or ID of the tenant to which the service principal belongs. - Type: string (or Expression with resultType string). - :type tenant: any - :param subscription_id: Data Lake Analytics account subscription ID (if different from Data - Factory account). Type: string (or Expression with resultType string). - :type subscription_id: any - :param resource_group_name: Data Lake Analytics account resource group name (if different from - Data Factory account). Type: string (or Expression with resultType string). - :type resource_group_name: any - :param data_lake_analytics_uri: Azure Data Lake Analytics URI Type: string (or Expression with - resultType string). - :type data_lake_analytics_uri: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'account_name': {'required': True}, - 'tenant': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'account_name': {'key': 'typeProperties.accountName', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'subscription_id': {'key': 'typeProperties.subscriptionId', 'type': 'object'}, - 'resource_group_name': {'key': 'typeProperties.resourceGroupName', 'type': 'object'}, - 'data_lake_analytics_uri': {'key': 'typeProperties.dataLakeAnalyticsUri', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - account_name: Any, - tenant: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - service_principal_id: Optional[Any] = None, - service_principal_key: Optional["SecretBase"] = None, - subscription_id: Optional[Any] = None, - resource_group_name: Optional[Any] = None, - data_lake_analytics_uri: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AzureDataLakeAnalyticsLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureDataLakeAnalytics' # type: str - self.account_name = account_name - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.subscription_id = subscription_id - self.resource_group_name = resource_group_name - self.data_lake_analytics_uri = data_lake_analytics_uri - self.encrypted_credential = encrypted_credential - - -class AzureDataLakeStoreDataset(Dataset): - """Azure Data Lake Store dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param folder_path: Path to the folder in the Azure Data Lake Store. Type: string (or - Expression with resultType string). - :type folder_path: any - :param file_name: The name of the file in the Azure Data Lake Store. Type: string (or - Expression with resultType string). - :type file_name: any - :param format: The format of the Data Lake Store. - :type format: ~azure.synapse.artifacts.v2020_12_01.models.DatasetStorageFormat - :param compression: The data compression method used for the item(s) in the Azure Data Lake - Store. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, - 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, - 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - folder_path: Optional[Any] = None, - file_name: Optional[Any] = None, - format: Optional["DatasetStorageFormat"] = None, - compression: Optional["DatasetCompression"] = None, - **kwargs - ): - super(AzureDataLakeStoreDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'AzureDataLakeStoreFile' # type: str - self.folder_path = folder_path - self.file_name = file_name - self.format = format - self.compression = compression - - -class AzureDataLakeStoreLinkedService(LinkedService): - """Azure Data Lake Store linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param data_lake_store_uri: Required. Data Lake Store service URI. Type: string (or Expression - with resultType string). - :type data_lake_store_uri: any - :param service_principal_id: The ID of the application used to authenticate against the Azure - Data Lake Store account. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: The Key of the application used to authenticate against the Azure - Data Lake Store account. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: The name or ID of the tenant to which the service principal belongs. Type: - string (or Expression with resultType string). - :type tenant: any - :param azure_cloud_type: Indicates the azure cloud type of the service principle auth. Allowed - values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data - factory regions’ cloud type. Type: string (or Expression with resultType string). - :type azure_cloud_type: any - :param account_name: Data Lake Store account name. Type: string (or Expression with resultType - string). - :type account_name: any - :param subscription_id: Data Lake Store account subscription ID (if different from Data Factory - account). Type: string (or Expression with resultType string). - :type subscription_id: any - :param resource_group_name: Data Lake Store account resource group name (if different from Data - Factory account). Type: string (or Expression with resultType string). - :type resource_group_name: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'data_lake_store_uri': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'data_lake_store_uri': {'key': 'typeProperties.dataLakeStoreUri', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'azure_cloud_type': {'key': 'typeProperties.azureCloudType', 'type': 'object'}, - 'account_name': {'key': 'typeProperties.accountName', 'type': 'object'}, - 'subscription_id': {'key': 'typeProperties.subscriptionId', 'type': 'object'}, - 'resource_group_name': {'key': 'typeProperties.resourceGroupName', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - data_lake_store_uri: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - service_principal_id: Optional[Any] = None, - service_principal_key: Optional["SecretBase"] = None, - tenant: Optional[Any] = None, - azure_cloud_type: Optional[Any] = None, - account_name: Optional[Any] = None, - subscription_id: Optional[Any] = None, - resource_group_name: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AzureDataLakeStoreLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureDataLakeStore' # type: str - self.data_lake_store_uri = data_lake_store_uri - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.azure_cloud_type = azure_cloud_type - self.account_name = account_name - self.subscription_id = subscription_id - self.resource_group_name = resource_group_name - self.encrypted_credential = encrypted_credential - - -class AzureDataLakeStoreLocation(DatasetLocation): - """The location of azure data lake store dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - folder_path: Optional[Any] = None, - file_name: Optional[Any] = None, - **kwargs - ): - super(AzureDataLakeStoreLocation, self).__init__(additional_properties=additional_properties, folder_path=folder_path, file_name=file_name, **kwargs) - self.type = 'AzureDataLakeStoreLocation' # type: str - - -class AzureDataLakeStoreReadSettings(StoreReadSettings): - """Azure data lake store read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: ADLS wildcardFolderPath. Type: string (or Expression with - resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: ADLS wildcardFileName. Type: string (or Expression with resultType - string). - :type wildcard_file_name: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param list_after: Lists files after the value (exclusive) based on file/folder names’ - lexicographical order. Applies under the folderPath in data set, and filter files/sub-folders - under the folderPath. Type: string (or Expression with resultType string). - :type list_after: any - :param list_before: Lists files before the value (inclusive) based on file/folder names’ - lexicographical order. Applies under the folderPath in data set, and filter files/sub-folders - under the folderPath. Type: string (or Expression with resultType string). - :type list_before: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'list_after': {'key': 'listAfter', 'type': 'object'}, - 'list_before': {'key': 'listBefore', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - 'modified_datetime_start': {'key': 'modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'modifiedDatetimeEnd', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - recursive: Optional[Any] = None, - wildcard_folder_path: Optional[Any] = None, - wildcard_file_name: Optional[Any] = None, - file_list_path: Optional[Any] = None, - list_after: Optional[Any] = None, - list_before: Optional[Any] = None, - enable_partition_discovery: Optional[bool] = None, - partition_root_path: Optional[Any] = None, - delete_files_after_completion: Optional[Any] = None, - modified_datetime_start: Optional[Any] = None, - modified_datetime_end: Optional[Any] = None, - **kwargs - ): - super(AzureDataLakeStoreReadSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AzureDataLakeStoreReadSettings' # type: str - self.recursive = recursive - self.wildcard_folder_path = wildcard_folder_path - self.wildcard_file_name = wildcard_file_name - self.file_list_path = file_list_path - self.list_after = list_after - self.list_before = list_before - self.enable_partition_discovery = enable_partition_discovery - self.partition_root_path = partition_root_path - self.delete_files_after_completion = delete_files_after_completion - self.modified_datetime_start = modified_datetime_start - self.modified_datetime_end = modified_datetime_end - - -class AzureDataLakeStoreSink(CopySink): - """A copy activity Azure Data Lake Store sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - :param enable_adls_single_file_parallel: Single File Parallel. - :type enable_adls_single_file_parallel: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - 'enable_adls_single_file_parallel': {'key': 'enableAdlsSingleFileParallel', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - copy_behavior: Optional[Any] = None, - enable_adls_single_file_parallel: Optional[Any] = None, - **kwargs - ): - super(AzureDataLakeStoreSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AzureDataLakeStoreSink' # type: str - self.copy_behavior = copy_behavior - self.enable_adls_single_file_parallel = enable_adls_single_file_parallel - - -class AzureDataLakeStoreSource(CopySource): - """A copy activity Azure Data Lake source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - recursive: Optional[Any] = None, - **kwargs - ): - super(AzureDataLakeStoreSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AzureDataLakeStoreSource' # type: str - self.recursive = recursive - - -class AzureDataLakeStoreWriteSettings(StoreWriteSettings): - """Azure data lake store write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - :param expiry_date_time: Specifies the expiry time of the written files. The time is applied to - the UTC time zone in the format of "2018-12-01T05:00:00Z". Default value is NULL. Type: integer - (or Expression with resultType integer). - :type expiry_date_time: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - 'expiry_date_time': {'key': 'expiryDateTime', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - copy_behavior: Optional[Any] = None, - expiry_date_time: Optional[Any] = None, - **kwargs - ): - super(AzureDataLakeStoreWriteSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) - self.type = 'AzureDataLakeStoreWriteSettings' # type: str - self.expiry_date_time = expiry_date_time - - -class Resource(msrest.serialization.Model): - """Common fields that are returned in the response for all Azure Resource Manager resources. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(Resource, self).__init__(**kwargs) - self.id = None - self.name = None - self.type = None - - -class AzureEntityResource(Resource): - """The resource model definition for an Azure Resource Manager resource with an etag. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(AzureEntityResource, self).__init__(**kwargs) - self.etag = None - - -class AzureFileStorageLinkedService(LinkedService): - """Azure File Storage linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. Host name of the server. Type: string (or Expression with resultType - string). - :type host: any - :param user_id: User ID to logon the server. Type: string (or Expression with resultType - string). - :type user_id: any - :param password: Password to logon the server. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param connection_string: The connection string. It is mutually exclusive with sasUri property. - Type: string, SecureString or AzureKeyVaultSecretReference. - :type connection_string: any - :param account_key: The Azure key vault secret reference of accountKey in connection string. - :type account_key: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param sas_uri: SAS URI of the Azure File resource. It is mutually exclusive with - connectionString property. Type: string, SecureString or AzureKeyVaultSecretReference. - :type sas_uri: any - :param sas_token: The Azure key vault secret reference of sasToken in sas uri. - :type sas_token: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param file_share: The azure file share name. It is required when auth with - accountKey/sasToken. Type: string (or Expression with resultType string). - :type file_share: any - :param snapshot: The azure file share snapshot version. Type: string (or Expression with - resultType string). - :type snapshot: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'user_id': {'key': 'typeProperties.userId', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'account_key': {'key': 'typeProperties.accountKey', 'type': 'AzureKeyVaultSecretReference'}, - 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'object'}, - 'sas_token': {'key': 'typeProperties.sasToken', 'type': 'AzureKeyVaultSecretReference'}, - 'file_share': {'key': 'typeProperties.fileShare', 'type': 'object'}, - 'snapshot': {'key': 'typeProperties.snapshot', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - host: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - user_id: Optional[Any] = None, - password: Optional["SecretBase"] = None, - connection_string: Optional[Any] = None, - account_key: Optional["AzureKeyVaultSecretReference"] = None, - sas_uri: Optional[Any] = None, - sas_token: Optional["AzureKeyVaultSecretReference"] = None, - file_share: Optional[Any] = None, - snapshot: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AzureFileStorageLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureFileStorage' # type: str - self.host = host - self.user_id = user_id - self.password = password - self.connection_string = connection_string - self.account_key = account_key - self.sas_uri = sas_uri - self.sas_token = sas_token - self.file_share = file_share - self.snapshot = snapshot - self.encrypted_credential = encrypted_credential - - -class AzureFileStorageLocation(DatasetLocation): - """The location of file server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - folder_path: Optional[Any] = None, - file_name: Optional[Any] = None, - **kwargs - ): - super(AzureFileStorageLocation, self).__init__(additional_properties=additional_properties, folder_path=folder_path, file_name=file_name, **kwargs) - self.type = 'AzureFileStorageLocation' # type: str - - -class AzureFileStorageReadSettings(StoreReadSettings): - """Azure File Storage read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: Azure File Storage wildcardFolderPath. Type: string (or Expression - with resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: Azure File Storage wildcardFileName. Type: string (or Expression - with resultType string). - :type wildcard_file_name: any - :param prefix: The prefix filter for the Azure File name starting from root path. Type: string - (or Expression with resultType string). - :type prefix: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'prefix': {'key': 'prefix', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - 'modified_datetime_start': {'key': 'modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'modifiedDatetimeEnd', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - recursive: Optional[Any] = None, - wildcard_folder_path: Optional[Any] = None, - wildcard_file_name: Optional[Any] = None, - prefix: Optional[Any] = None, - file_list_path: Optional[Any] = None, - enable_partition_discovery: Optional[bool] = None, - partition_root_path: Optional[Any] = None, - delete_files_after_completion: Optional[Any] = None, - modified_datetime_start: Optional[Any] = None, - modified_datetime_end: Optional[Any] = None, - **kwargs - ): - super(AzureFileStorageReadSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AzureFileStorageReadSettings' # type: str - self.recursive = recursive - self.wildcard_folder_path = wildcard_folder_path - self.wildcard_file_name = wildcard_file_name - self.prefix = prefix - self.file_list_path = file_list_path - self.enable_partition_discovery = enable_partition_discovery - self.partition_root_path = partition_root_path - self.delete_files_after_completion = delete_files_after_completion - self.modified_datetime_start = modified_datetime_start - self.modified_datetime_end = modified_datetime_end - - -class AzureFileStorageWriteSettings(StoreWriteSettings): - """Azure File Storage write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - copy_behavior: Optional[Any] = None, - **kwargs - ): - super(AzureFileStorageWriteSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) - self.type = 'AzureFileStorageWriteSettings' # type: str - - -class AzureFunctionActivity(ExecutionActivity): - """Azure Function activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param method: Required. Rest API method for target endpoint. Possible values include: "GET", - "POST", "PUT", "DELETE", "OPTIONS", "HEAD", "TRACE". - :type method: str or ~azure.synapse.artifacts.v2020_12_01.models.AzureFunctionActivityMethod - :param function_name: Required. Name of the Function that the Azure Function Activity will - call. Type: string (or Expression with resultType string). - :type function_name: any - :param headers: Represents the headers that will be sent to the request. For example, to set - the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": - "application/json" }. Type: string (or Expression with resultType string). - :type headers: any - :param body: Represents the payload that will be sent to the endpoint. Required for POST/PUT - method, not allowed for GET method Type: string (or Expression with resultType string). - :type body: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'method': {'required': True}, - 'function_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'method': {'key': 'typeProperties.method', 'type': 'str'}, - 'function_name': {'key': 'typeProperties.functionName', 'type': 'object'}, - 'headers': {'key': 'typeProperties.headers', 'type': 'object'}, - 'body': {'key': 'typeProperties.body', 'type': 'object'}, - } - - def __init__( - self, - *, - name: str, - method: Union[str, "AzureFunctionActivityMethod"], - function_name: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - headers: Optional[Any] = None, - body: Optional[Any] = None, - **kwargs - ): - super(AzureFunctionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'AzureFunctionActivity' # type: str - self.method = method - self.function_name = function_name - self.headers = headers - self.body = body - - -class AzureFunctionLinkedService(LinkedService): - """Azure Function linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param function_app_url: Required. The endpoint of the Azure Function App. URL will be in the - format https://:code:``.azurewebsites.net. - :type function_app_url: any - :param function_key: Function or Host key for Azure Function App. - :type function_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'function_app_url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'function_app_url': {'key': 'typeProperties.functionAppUrl', 'type': 'object'}, - 'function_key': {'key': 'typeProperties.functionKey', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - function_app_url: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - function_key: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AzureFunctionLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureFunction' # type: str - self.function_app_url = function_app_url - self.function_key = function_key - self.encrypted_credential = encrypted_credential - - -class AzureKeyVaultLinkedService(LinkedService): - """Azure Key Vault linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param base_url: Required. The base URL of the Azure Key Vault. e.g. - https://myakv.vault.azure.net Type: string (or Expression with resultType string). - :type base_url: any - """ - - _validation = { - 'type': {'required': True}, - 'base_url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'base_url': {'key': 'typeProperties.baseUrl', 'type': 'object'}, - } - - def __init__( - self, - *, - base_url: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - **kwargs - ): - super(AzureKeyVaultLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureKeyVault' # type: str - self.base_url = base_url - - -class SecretBase(msrest.serialization.Model): - """The base definition of a secret type. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AzureKeyVaultSecretReference, SecureString. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Type of the secret.Constant filled by server. - :type type: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - } - - _subtype_map = { - 'type': {'AzureKeyVaultSecret': 'AzureKeyVaultSecretReference', 'SecureString': 'SecureString'} - } - - def __init__( - self, - **kwargs - ): - super(SecretBase, self).__init__(**kwargs) - self.type = None # type: Optional[str] - - -class AzureKeyVaultSecretReference(SecretBase): - """Azure Key Vault secret reference. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Type of the secret.Constant filled by server. - :type type: str - :param store: Required. The Azure Key Vault linked service reference. - :type store: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param secret_name: Required. The name of the secret in Azure Key Vault. Type: string (or - Expression with resultType string). - :type secret_name: any - :param secret_version: The version of the secret in Azure Key Vault. The default value is the - latest version of the secret. Type: string (or Expression with resultType string). - :type secret_version: any - """ - - _validation = { - 'type': {'required': True}, - 'store': {'required': True}, - 'secret_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'store': {'key': 'store', 'type': 'LinkedServiceReference'}, - 'secret_name': {'key': 'secretName', 'type': 'object'}, - 'secret_version': {'key': 'secretVersion', 'type': 'object'}, - } - - def __init__( - self, - *, - store: "LinkedServiceReference", - secret_name: Any, - secret_version: Optional[Any] = None, - **kwargs - ): - super(AzureKeyVaultSecretReference, self).__init__(**kwargs) - self.type = 'AzureKeyVaultSecret' # type: str - self.store = store - self.secret_name = secret_name - self.secret_version = secret_version - - -class AzureMariaDBLinkedService(LinkedService): - """Azure Database for MariaDB linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: An ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_string: Optional[Any] = None, - pwd: Optional["AzureKeyVaultSecretReference"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AzureMariaDBLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureMariaDB' # type: str - self.connection_string = connection_string - self.pwd = pwd - self.encrypted_credential = encrypted_credential - - -class AzureMariaDBSource(TabularSource): - """A copy activity Azure MariaDB source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(AzureMariaDBSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'AzureMariaDBSource' # type: str - self.query = query - - -class AzureMariaDBTableDataset(Dataset): - """Azure Database for MariaDB dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(AzureMariaDBTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'AzureMariaDBTable' # type: str - self.table_name = table_name - - -class AzureMLBatchExecutionActivity(ExecutionActivity): - """Azure ML Batch Execution activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param global_parameters: Key,Value pairs to be passed to the Azure ML Batch Execution Service - endpoint. Keys must match the names of web service parameters defined in the published Azure ML - web service. Values will be passed in the GlobalParameters property of the Azure ML batch - execution request. - :type global_parameters: dict[str, any] - :param web_service_outputs: Key,Value pairs, mapping the names of Azure ML endpoint's Web - Service Outputs to AzureMLWebServiceFile objects specifying the output Blob locations. This - information will be passed in the WebServiceOutputs property of the Azure ML batch execution - request. - :type web_service_outputs: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.AzureMLWebServiceFile] - :param web_service_inputs: Key,Value pairs, mapping the names of Azure ML endpoint's Web - Service Inputs to AzureMLWebServiceFile objects specifying the input Blob locations.. This - information will be passed in the WebServiceInputs property of the Azure ML batch execution - request. - :type web_service_inputs: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.AzureMLWebServiceFile] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'global_parameters': {'key': 'typeProperties.globalParameters', 'type': '{object}'}, - 'web_service_outputs': {'key': 'typeProperties.webServiceOutputs', 'type': '{AzureMLWebServiceFile}'}, - 'web_service_inputs': {'key': 'typeProperties.webServiceInputs', 'type': '{AzureMLWebServiceFile}'}, - } - - def __init__( - self, - *, - name: str, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - global_parameters: Optional[Dict[str, Any]] = None, - web_service_outputs: Optional[Dict[str, "AzureMLWebServiceFile"]] = None, - web_service_inputs: Optional[Dict[str, "AzureMLWebServiceFile"]] = None, - **kwargs - ): - super(AzureMLBatchExecutionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'AzureMLBatchExecution' # type: str - self.global_parameters = global_parameters - self.web_service_outputs = web_service_outputs - self.web_service_inputs = web_service_inputs - - -class AzureMLExecutePipelineActivity(ExecutionActivity): - """Azure ML Execute Pipeline activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param ml_pipeline_id: Required. ID of the published Azure ML pipeline. Type: string (or - Expression with resultType string). - :type ml_pipeline_id: any - :param experiment_name: Run history experiment name of the pipeline run. This information will - be passed in the ExperimentName property of the published pipeline execution request. Type: - string (or Expression with resultType string). - :type experiment_name: any - :param ml_pipeline_parameters: Key,Value pairs to be passed to the published Azure ML pipeline - endpoint. Keys must match the names of pipeline parameters defined in the published pipeline. - Values will be passed in the ParameterAssignments property of the published pipeline execution - request. Type: object with key value pairs (or Expression with resultType object). - :type ml_pipeline_parameters: any - :param ml_parent_run_id: The parent Azure ML Service pipeline run id. This information will be - passed in the ParentRunId property of the published pipeline execution request. Type: string - (or Expression with resultType string). - :type ml_parent_run_id: any - :param continue_on_step_failure: Whether to continue execution of other steps in the - PipelineRun if a step fails. This information will be passed in the continueOnStepFailure - property of the published pipeline execution request. Type: boolean (or Expression with - resultType boolean). - :type continue_on_step_failure: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'ml_pipeline_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'ml_pipeline_id': {'key': 'typeProperties.mlPipelineId', 'type': 'object'}, - 'experiment_name': {'key': 'typeProperties.experimentName', 'type': 'object'}, - 'ml_pipeline_parameters': {'key': 'typeProperties.mlPipelineParameters', 'type': 'object'}, - 'ml_parent_run_id': {'key': 'typeProperties.mlParentRunId', 'type': 'object'}, - 'continue_on_step_failure': {'key': 'typeProperties.continueOnStepFailure', 'type': 'object'}, - } - - def __init__( - self, - *, - name: str, - ml_pipeline_id: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - experiment_name: Optional[Any] = None, - ml_pipeline_parameters: Optional[Any] = None, - ml_parent_run_id: Optional[Any] = None, - continue_on_step_failure: Optional[Any] = None, - **kwargs - ): - super(AzureMLExecutePipelineActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'AzureMLExecutePipeline' # type: str - self.ml_pipeline_id = ml_pipeline_id - self.experiment_name = experiment_name - self.ml_pipeline_parameters = ml_pipeline_parameters - self.ml_parent_run_id = ml_parent_run_id - self.continue_on_step_failure = continue_on_step_failure - - -class AzureMLLinkedService(LinkedService): - """Azure ML Studio Web Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param ml_endpoint: Required. The Batch Execution REST URL for an Azure ML Studio Web Service - endpoint. Type: string (or Expression with resultType string). - :type ml_endpoint: any - :param api_key: Required. The API key for accessing the Azure ML model endpoint. - :type api_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param update_resource_endpoint: The Update Resource REST URL for an Azure ML Studio Web - Service endpoint. Type: string (or Expression with resultType string). - :type update_resource_endpoint: any - :param service_principal_id: The ID of the service principal used to authenticate against the - ARM-based updateResourceEndpoint of an Azure ML Studio web service. Type: string (or Expression - with resultType string). - :type service_principal_id: any - :param service_principal_key: The key of the service principal used to authenticate against the - ARM-based updateResourceEndpoint of an Azure ML Studio web service. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: The name or ID of the tenant to which the service principal belongs. Type: - string (or Expression with resultType string). - :type tenant: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'ml_endpoint': {'required': True}, - 'api_key': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'ml_endpoint': {'key': 'typeProperties.mlEndpoint', 'type': 'object'}, - 'api_key': {'key': 'typeProperties.apiKey', 'type': 'SecretBase'}, - 'update_resource_endpoint': {'key': 'typeProperties.updateResourceEndpoint', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - ml_endpoint: Any, - api_key: "SecretBase", - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - update_resource_endpoint: Optional[Any] = None, - service_principal_id: Optional[Any] = None, - service_principal_key: Optional["SecretBase"] = None, - tenant: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AzureMLLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureML' # type: str - self.ml_endpoint = ml_endpoint - self.api_key = api_key - self.update_resource_endpoint = update_resource_endpoint - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.encrypted_credential = encrypted_credential - - -class AzureMLServiceLinkedService(LinkedService): - """Azure ML Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param subscription_id: Required. Azure ML Service workspace subscription ID. Type: string (or - Expression with resultType string). - :type subscription_id: any - :param resource_group_name: Required. Azure ML Service workspace resource group name. Type: - string (or Expression with resultType string). - :type resource_group_name: any - :param ml_workspace_name: Required. Azure ML Service workspace name. Type: string (or - Expression with resultType string). - :type ml_workspace_name: any - :param service_principal_id: The ID of the service principal used to authenticate against the - endpoint of a published Azure ML Service pipeline. Type: string (or Expression with resultType - string). - :type service_principal_id: any - :param service_principal_key: The key of the service principal used to authenticate against the - endpoint of a published Azure ML Service pipeline. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: The name or ID of the tenant to which the service principal belongs. Type: - string (or Expression with resultType string). - :type tenant: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'subscription_id': {'required': True}, - 'resource_group_name': {'required': True}, - 'ml_workspace_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'subscription_id': {'key': 'typeProperties.subscriptionId', 'type': 'object'}, - 'resource_group_name': {'key': 'typeProperties.resourceGroupName', 'type': 'object'}, - 'ml_workspace_name': {'key': 'typeProperties.mlWorkspaceName', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - subscription_id: Any, - resource_group_name: Any, - ml_workspace_name: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - service_principal_id: Optional[Any] = None, - service_principal_key: Optional["SecretBase"] = None, - tenant: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AzureMLServiceLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureMLService' # type: str - self.subscription_id = subscription_id - self.resource_group_name = resource_group_name - self.ml_workspace_name = ml_workspace_name - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.encrypted_credential = encrypted_credential - - -class AzureMLUpdateResourceActivity(ExecutionActivity): - """Azure ML Update Resource management activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param trained_model_name: Required. Name of the Trained Model module in the Web Service - experiment to be updated. Type: string (or Expression with resultType string). - :type trained_model_name: any - :param trained_model_linked_service_name: Required. Name of Azure Storage linked service - holding the .ilearner file that will be uploaded by the update operation. - :type trained_model_linked_service_name: - ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param trained_model_file_path: Required. The relative file path in trainedModelLinkedService - to represent the .ilearner file that will be uploaded by the update operation. Type: string - (or Expression with resultType string). - :type trained_model_file_path: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'trained_model_name': {'required': True}, - 'trained_model_linked_service_name': {'required': True}, - 'trained_model_file_path': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'trained_model_name': {'key': 'typeProperties.trainedModelName', 'type': 'object'}, - 'trained_model_linked_service_name': {'key': 'typeProperties.trainedModelLinkedServiceName', 'type': 'LinkedServiceReference'}, - 'trained_model_file_path': {'key': 'typeProperties.trainedModelFilePath', 'type': 'object'}, - } - - def __init__( - self, - *, - name: str, - trained_model_name: Any, - trained_model_linked_service_name: "LinkedServiceReference", - trained_model_file_path: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - **kwargs - ): - super(AzureMLUpdateResourceActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'AzureMLUpdateResource' # type: str - self.trained_model_name = trained_model_name - self.trained_model_linked_service_name = trained_model_linked_service_name - self.trained_model_file_path = trained_model_file_path - - -class AzureMLWebServiceFile(msrest.serialization.Model): - """Azure ML WebService Input/Output file. - - All required parameters must be populated in order to send to Azure. - - :param file_path: Required. The relative file path, including container name, in the Azure Blob - Storage specified by the LinkedService. Type: string (or Expression with resultType string). - :type file_path: any - :param linked_service_name: Required. Reference to an Azure Storage LinkedService, where Azure - ML WebService Input/Output file located. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - """ - - _validation = { - 'file_path': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'file_path': {'key': 'filePath', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - } - - def __init__( - self, - *, - file_path: Any, - linked_service_name: "LinkedServiceReference", - **kwargs - ): - super(AzureMLWebServiceFile, self).__init__(**kwargs) - self.file_path = file_path - self.linked_service_name = linked_service_name - - -class AzureMySqlLinkedService(LinkedService): - """Azure MySQL database linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param password: The Azure key vault secret reference of password in connection string. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - connection_string: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - password: Optional["AzureKeyVaultSecretReference"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AzureMySqlLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureMySql' # type: str - self.connection_string = connection_string - self.password = password - self.encrypted_credential = encrypted_credential - - -class AzureMySqlSink(CopySink): - """A copy activity Azure MySql sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param pre_copy_script: A query to execute before starting the copy. Type: string (or - Expression with resultType string). - :type pre_copy_script: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - pre_copy_script: Optional[Any] = None, - **kwargs - ): - super(AzureMySqlSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AzureMySqlSink' # type: str - self.pre_copy_script = pre_copy_script - - -class AzureMySqlSource(TabularSource): - """A copy activity Azure MySQL source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(AzureMySqlSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'AzureMySqlSource' # type: str - self.query = query - - -class AzureMySqlTableDataset(Dataset): - """The Azure MySQL database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The Azure MySQL database table name. Type: string (or Expression with - resultType string). - :type table_name: any - :param table: The name of Azure MySQL database table. Type: string (or Expression with - resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - table: Optional[Any] = None, - **kwargs - ): - super(AzureMySqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'AzureMySqlTable' # type: str - self.table_name = table_name - self.table = table - - -class AzurePostgreSqlLinkedService(LinkedService): - """Azure PostgreSQL linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: An ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param password: The Azure key vault secret reference of password in connection string. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_string: Optional[Any] = None, - password: Optional["AzureKeyVaultSecretReference"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AzurePostgreSqlLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzurePostgreSql' # type: str - self.connection_string = connection_string - self.password = password - self.encrypted_credential = encrypted_credential - - -class AzurePostgreSqlSink(CopySink): - """A copy activity Azure PostgreSQL sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param pre_copy_script: A query to execute before starting the copy. Type: string (or - Expression with resultType string). - :type pre_copy_script: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - pre_copy_script: Optional[Any] = None, - **kwargs - ): - super(AzurePostgreSqlSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AzurePostgreSqlSink' # type: str - self.pre_copy_script = pre_copy_script - - -class AzurePostgreSqlSource(TabularSource): - """A copy activity Azure PostgreSQL source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(AzurePostgreSqlSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'AzurePostgreSqlSource' # type: str - self.query = query - - -class AzurePostgreSqlTableDataset(Dataset): - """Azure PostgreSQL dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name of the Azure PostgreSQL database which includes both schema - and table. Type: string (or Expression with resultType string). - :type table_name: any - :param table: The table name of the Azure PostgreSQL database. Type: string (or Expression with - resultType string). - :type table: any - :param schema_type_properties_schema: The schema name of the Azure PostgreSQL database. Type: - string (or Expression with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - table: Optional[Any] = None, - schema_type_properties_schema: Optional[Any] = None, - **kwargs - ): - super(AzurePostgreSqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'AzurePostgreSqlTable' # type: str - self.table_name = table_name - self.table = table - self.schema_type_properties_schema = schema_type_properties_schema - - -class AzureQueueSink(CopySink): - """A copy activity Azure Queue sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - **kwargs - ): - super(AzureQueueSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AzureQueueSink' # type: str - - -class AzureSearchIndexDataset(Dataset): - """The Azure Search Index. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param index_name: Required. The name of the Azure Search Index. Type: string (or Expression - with resultType string). - :type index_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'index_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'index_name': {'key': 'typeProperties.indexName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - index_name: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - **kwargs - ): - super(AzureSearchIndexDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'AzureSearchIndex' # type: str - self.index_name = index_name - - -class AzureSearchIndexSink(CopySink): - """A copy activity Azure Search Index sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param write_behavior: Specify the write behavior when upserting documents into Azure Search - Index. Possible values include: "Merge", "Upload". - :type write_behavior: str or - ~azure.synapse.artifacts.v2020_12_01.models.AzureSearchIndexWriteBehaviorType - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - write_behavior: Optional[Union[str, "AzureSearchIndexWriteBehaviorType"]] = None, - **kwargs - ): - super(AzureSearchIndexSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AzureSearchIndexSink' # type: str - self.write_behavior = write_behavior - - -class AzureSearchLinkedService(LinkedService): - """Linked service for Windows Azure Search Service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param url: Required. URL for Azure Search service. Type: string (or Expression with resultType - string). - :type url: any - :param key: Admin Key for Azure Search service. - :type key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'key': {'key': 'typeProperties.key', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - url: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - key: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AzureSearchLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureSearch' # type: str - self.url = url - self.key = key - self.encrypted_credential = encrypted_credential - - -class AzureSqlDatabaseLinkedService(LinkedService): - """Microsoft Azure SQL Database linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param password: The Azure key vault secret reference of password in connection string. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param service_principal_id: The ID of the service principal used to authenticate against Azure - SQL Database. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: The key of the service principal used to authenticate against - Azure SQL Database. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: The name or ID of the tenant to which the service principal belongs. Type: - string (or Expression with resultType string). - :type tenant: any - :param azure_cloud_type: Indicates the azure cloud type of the service principle auth. Allowed - values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data - factory regions’ cloud type. Type: string (or Expression with resultType string). - :type azure_cloud_type: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'azure_cloud_type': {'key': 'typeProperties.azureCloudType', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - connection_string: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - password: Optional["AzureKeyVaultSecretReference"] = None, - service_principal_id: Optional[Any] = None, - service_principal_key: Optional["SecretBase"] = None, - tenant: Optional[Any] = None, - azure_cloud_type: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AzureSqlDatabaseLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureSqlDatabase' # type: str - self.connection_string = connection_string - self.password = password - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.azure_cloud_type = azure_cloud_type - self.encrypted_credential = encrypted_credential - - -class AzureSqlDWLinkedService(LinkedService): - """Azure SQL Data Warehouse linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. Type: string, SecureString or AzureKeyVaultSecretReference. - :type connection_string: any - :param password: The Azure key vault secret reference of password in connection string. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param service_principal_id: The ID of the service principal used to authenticate against Azure - SQL Data Warehouse. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: The key of the service principal used to authenticate against - Azure SQL Data Warehouse. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: The name or ID of the tenant to which the service principal belongs. Type: - string (or Expression with resultType string). - :type tenant: any - :param azure_cloud_type: Indicates the azure cloud type of the service principle auth. Allowed - values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data - factory regions’ cloud type. Type: string (or Expression with resultType string). - :type azure_cloud_type: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'azure_cloud_type': {'key': 'typeProperties.azureCloudType', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - connection_string: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - password: Optional["AzureKeyVaultSecretReference"] = None, - service_principal_id: Optional[Any] = None, - service_principal_key: Optional["SecretBase"] = None, - tenant: Optional[Any] = None, - azure_cloud_type: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AzureSqlDWLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureSqlDW' # type: str - self.connection_string = connection_string - self.password = password - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.azure_cloud_type = azure_cloud_type - self.encrypted_credential = encrypted_credential - - -class AzureSqlDWTableDataset(Dataset): - """The Azure SQL Data Warehouse dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param schema_type_properties_schema: The schema name of the Azure SQL Data Warehouse. Type: - string (or Expression with resultType string). - :type schema_type_properties_schema: any - :param table: The table name of the Azure SQL Data Warehouse. Type: string (or Expression with - resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - schema_type_properties_schema: Optional[Any] = None, - table: Optional[Any] = None, - **kwargs - ): - super(AzureSqlDWTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'AzureSqlDWTable' # type: str - self.table_name = table_name - self.schema_type_properties_schema = schema_type_properties_schema - self.table = table - - -class AzureSqlMILinkedService(LinkedService): - """Azure SQL Managed Instance linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param password: The Azure key vault secret reference of password in connection string. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param service_principal_id: The ID of the service principal used to authenticate against Azure - SQL Managed Instance. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: The key of the service principal used to authenticate against - Azure SQL Managed Instance. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: The name or ID of the tenant to which the service principal belongs. Type: - string (or Expression with resultType string). - :type tenant: any - :param azure_cloud_type: Indicates the azure cloud type of the service principle auth. Allowed - values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data - factory regions’ cloud type. Type: string (or Expression with resultType string). - :type azure_cloud_type: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'azure_cloud_type': {'key': 'typeProperties.azureCloudType', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - connection_string: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - password: Optional["AzureKeyVaultSecretReference"] = None, - service_principal_id: Optional[Any] = None, - service_principal_key: Optional["SecretBase"] = None, - tenant: Optional[Any] = None, - azure_cloud_type: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(AzureSqlMILinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureSqlMI' # type: str - self.connection_string = connection_string - self.password = password - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.azure_cloud_type = azure_cloud_type - self.encrypted_credential = encrypted_credential - - -class AzureSqlMITableDataset(Dataset): - """The Azure SQL Managed Instance dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param schema_type_properties_schema: The schema name of the Azure SQL Managed Instance. Type: - string (or Expression with resultType string). - :type schema_type_properties_schema: any - :param table: The table name of the Azure SQL Managed Instance dataset. Type: string (or - Expression with resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - schema_type_properties_schema: Optional[Any] = None, - table: Optional[Any] = None, - **kwargs - ): - super(AzureSqlMITableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'AzureSqlMITable' # type: str - self.table_name = table_name - self.schema_type_properties_schema = schema_type_properties_schema - self.table = table - - -class AzureSqlSink(CopySink): - """A copy activity Azure SQL sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param sql_writer_stored_procedure_name: SQL writer stored procedure name. Type: string (or - Expression with resultType string). - :type sql_writer_stored_procedure_name: any - :param sql_writer_table_type: SQL writer table type. Type: string (or Expression with - resultType string). - :type sql_writer_table_type: any - :param pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType - string). - :type pre_copy_script: any - :param stored_procedure_parameters: SQL stored procedure parameters. - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - :param stored_procedure_table_type_parameter_name: The stored procedure parameter name of the - table type. Type: string (or Expression with resultType string). - :type stored_procedure_table_type_parameter_name: any - :param table_option: The option to handle sink table, such as autoCreate. For now only - 'autoCreate' value is supported. Type: string (or Expression with resultType string). - :type table_option: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'sql_writer_stored_procedure_name': {'key': 'sqlWriterStoredProcedureName', 'type': 'object'}, - 'sql_writer_table_type': {'key': 'sqlWriterTableType', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - 'stored_procedure_table_type_parameter_name': {'key': 'storedProcedureTableTypeParameterName', 'type': 'object'}, - 'table_option': {'key': 'tableOption', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - sql_writer_stored_procedure_name: Optional[Any] = None, - sql_writer_table_type: Optional[Any] = None, - pre_copy_script: Optional[Any] = None, - stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, - stored_procedure_table_type_parameter_name: Optional[Any] = None, - table_option: Optional[Any] = None, - **kwargs - ): - super(AzureSqlSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AzureSqlSink' # type: str - self.sql_writer_stored_procedure_name = sql_writer_stored_procedure_name - self.sql_writer_table_type = sql_writer_table_type - self.pre_copy_script = pre_copy_script - self.stored_procedure_parameters = stored_procedure_parameters - self.stored_procedure_table_type_parameter_name = stored_procedure_table_type_parameter_name - self.table_option = table_option - - -class AzureSqlSource(TabularSource): - """A copy activity Azure SQL source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). - :type sql_reader_query: any - :param sql_reader_stored_procedure_name: Name of the stored procedure for a SQL Database - source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression - with resultType string). - :type sql_reader_stored_procedure_name: any - :param stored_procedure_parameters: Value and type setting for stored procedure parameters. - Example: "{Parameter1: {value: "1", type: "int"}}". - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - :param produce_additional_types: Which additional types to produce. - :type produce_additional_types: any - :param partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". - :type partition_option: any - :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.SqlPartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, - 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - 'produce_additional_types': {'key': 'produceAdditionalTypes', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'object'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - sql_reader_query: Optional[Any] = None, - sql_reader_stored_procedure_name: Optional[Any] = None, - stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, - produce_additional_types: Optional[Any] = None, - partition_option: Optional[Any] = None, - partition_settings: Optional["SqlPartitionSettings"] = None, - **kwargs - ): - super(AzureSqlSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'AzureSqlSource' # type: str - self.sql_reader_query = sql_reader_query - self.sql_reader_stored_procedure_name = sql_reader_stored_procedure_name - self.stored_procedure_parameters = stored_procedure_parameters - self.produce_additional_types = produce_additional_types - self.partition_option = partition_option - self.partition_settings = partition_settings - - -class AzureSqlTableDataset(Dataset): - """The Azure SQL Server database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param schema_type_properties_schema: The schema name of the Azure SQL database. Type: string - (or Expression with resultType string). - :type schema_type_properties_schema: any - :param table: The table name of the Azure SQL database. Type: string (or Expression with - resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - schema_type_properties_schema: Optional[Any] = None, - table: Optional[Any] = None, - **kwargs - ): - super(AzureSqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'AzureSqlTable' # type: str - self.table_name = table_name - self.schema_type_properties_schema = schema_type_properties_schema - self.table = table - - -class AzureStorageLinkedService(LinkedService): - """The storage account linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: The connection string. It is mutually exclusive with sasUri property. - Type: string, SecureString or AzureKeyVaultSecretReference. - :type connection_string: any - :param account_key: The Azure key vault secret reference of accountKey in connection string. - :type account_key: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param sas_uri: SAS URI of the Azure Storage resource. It is mutually exclusive with - connectionString property. Type: string, SecureString or AzureKeyVaultSecretReference. - :type sas_uri: any - :param sas_token: The Azure key vault secret reference of sasToken in sas uri. - :type sas_token: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'account_key': {'key': 'typeProperties.accountKey', 'type': 'AzureKeyVaultSecretReference'}, - 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'object'}, - 'sas_token': {'key': 'typeProperties.sasToken', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_string: Optional[Any] = None, - account_key: Optional["AzureKeyVaultSecretReference"] = None, - sas_uri: Optional[Any] = None, - sas_token: Optional["AzureKeyVaultSecretReference"] = None, - encrypted_credential: Optional[str] = None, - **kwargs - ): - super(AzureStorageLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureStorage' # type: str - self.connection_string = connection_string - self.account_key = account_key - self.sas_uri = sas_uri - self.sas_token = sas_token - self.encrypted_credential = encrypted_credential - - -class AzureTableDataset(Dataset): - """The Azure Table storage dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: Required. The table name of the Azure Table storage. Type: string (or - Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'table_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - table_name: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - **kwargs - ): - super(AzureTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'AzureTable' # type: str - self.table_name = table_name - - -class AzureTableSink(CopySink): - """A copy activity Azure Table sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param azure_table_default_partition_key_value: Azure Table default partition key value. Type: - string (or Expression with resultType string). - :type azure_table_default_partition_key_value: any - :param azure_table_partition_key_name: Azure Table partition key name. Type: string (or - Expression with resultType string). - :type azure_table_partition_key_name: any - :param azure_table_row_key_name: Azure Table row key name. Type: string (or Expression with - resultType string). - :type azure_table_row_key_name: any - :param azure_table_insert_type: Azure Table insert type. Type: string (or Expression with - resultType string). - :type azure_table_insert_type: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'azure_table_default_partition_key_value': {'key': 'azureTableDefaultPartitionKeyValue', 'type': 'object'}, - 'azure_table_partition_key_name': {'key': 'azureTablePartitionKeyName', 'type': 'object'}, - 'azure_table_row_key_name': {'key': 'azureTableRowKeyName', 'type': 'object'}, - 'azure_table_insert_type': {'key': 'azureTableInsertType', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - azure_table_default_partition_key_value: Optional[Any] = None, - azure_table_partition_key_name: Optional[Any] = None, - azure_table_row_key_name: Optional[Any] = None, - azure_table_insert_type: Optional[Any] = None, - **kwargs - ): - super(AzureTableSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'AzureTableSink' # type: str - self.azure_table_default_partition_key_value = azure_table_default_partition_key_value - self.azure_table_partition_key_name = azure_table_partition_key_name - self.azure_table_row_key_name = azure_table_row_key_name - self.azure_table_insert_type = azure_table_insert_type - - -class AzureTableSource(TabularSource): - """A copy activity Azure Table source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param azure_table_source_query: Azure Table source query. Type: string (or Expression with - resultType string). - :type azure_table_source_query: any - :param azure_table_source_ignore_table_not_found: Azure Table source ignore table not found. - Type: boolean (or Expression with resultType boolean). - :type azure_table_source_ignore_table_not_found: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'azure_table_source_query': {'key': 'azureTableSourceQuery', 'type': 'object'}, - 'azure_table_source_ignore_table_not_found': {'key': 'azureTableSourceIgnoreTableNotFound', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - azure_table_source_query: Optional[Any] = None, - azure_table_source_ignore_table_not_found: Optional[Any] = None, - **kwargs - ): - super(AzureTableSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'AzureTableSource' # type: str - self.azure_table_source_query = azure_table_source_query - self.azure_table_source_ignore_table_not_found = azure_table_source_ignore_table_not_found - - -class AzureTableStorageLinkedService(LinkedService): - """The azure table storage linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: The connection string. It is mutually exclusive with sasUri property. - Type: string, SecureString or AzureKeyVaultSecretReference. - :type connection_string: any - :param account_key: The Azure key vault secret reference of accountKey in connection string. - :type account_key: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param sas_uri: SAS URI of the Azure Storage resource. It is mutually exclusive with - connectionString property. Type: string, SecureString or AzureKeyVaultSecretReference. - :type sas_uri: any - :param sas_token: The Azure key vault secret reference of sasToken in sas uri. - :type sas_token: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'account_key': {'key': 'typeProperties.accountKey', 'type': 'AzureKeyVaultSecretReference'}, - 'sas_uri': {'key': 'typeProperties.sasUri', 'type': 'object'}, - 'sas_token': {'key': 'typeProperties.sasToken', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_string: Optional[Any] = None, - account_key: Optional["AzureKeyVaultSecretReference"] = None, - sas_uri: Optional[Any] = None, - sas_token: Optional["AzureKeyVaultSecretReference"] = None, - encrypted_credential: Optional[str] = None, - **kwargs - ): - super(AzureTableStorageLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'AzureTableStorage' # type: str - self.connection_string = connection_string - self.account_key = account_key - self.sas_uri = sas_uri - self.sas_token = sas_token - self.encrypted_credential = encrypted_credential - - -class BigDataPoolReference(msrest.serialization.Model): - """Big data pool reference. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Big data pool reference type. Possible values include: - "BigDataPoolReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.BigDataPoolReferenceType - :param reference_name: Required. Reference big data pool name. - :type reference_name: str - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - } - - def __init__( - self, - *, - type: Union[str, "BigDataPoolReferenceType"], - reference_name: str, - **kwargs - ): - super(BigDataPoolReference, self).__init__(**kwargs) - self.type = type - self.reference_name = reference_name - - -class TrackedResource(Resource): - """The resource model definition for an Azure Resource Manager tracked top level resource which has 'tags' and a 'location'. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :param tags: A set of tags. Resource tags. - :type tags: dict[str, str] - :param location: Required. The geo-location where the resource lives. - :type location: str - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'location': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'tags': {'key': 'tags', 'type': '{str}'}, - 'location': {'key': 'location', 'type': 'str'}, - } - - def __init__( - self, - *, - location: str, - tags: Optional[Dict[str, str]] = None, - **kwargs - ): - super(TrackedResource, self).__init__(**kwargs) - self.tags = tags - self.location = location - - -class BigDataPoolResourceInfo(TrackedResource): - """A Big Data pool. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :param tags: A set of tags. Resource tags. - :type tags: dict[str, str] - :param location: Required. The geo-location where the resource lives. - :type location: str - :param provisioning_state: The state of the Big Data pool. - :type provisioning_state: str - :param auto_scale: Auto-scaling properties. - :type auto_scale: ~azure.synapse.artifacts.v2020_12_01.models.AutoScaleProperties - :param creation_date: The time when the Big Data pool was created. - :type creation_date: ~datetime.datetime - :param auto_pause: Auto-pausing properties. - :type auto_pause: ~azure.synapse.artifacts.v2020_12_01.models.AutoPauseProperties - :param is_compute_isolation_enabled: Whether compute isolation is required or not. - :type is_compute_isolation_enabled: bool - :param session_level_packages_enabled: Whether session level packages enabled. - :type session_level_packages_enabled: bool - :param cache_size: The cache size. - :type cache_size: int - :param dynamic_executor_allocation: Dynamic Executor Allocation. - :type dynamic_executor_allocation: - ~azure.synapse.artifacts.v2020_12_01.models.DynamicExecutorAllocation - :param spark_events_folder: The Spark events folder. - :type spark_events_folder: str - :param node_count: The number of nodes in the Big Data pool. - :type node_count: int - :param library_requirements: Library version requirements. - :type library_requirements: ~azure.synapse.artifacts.v2020_12_01.models.LibraryRequirements - :param custom_libraries: List of custom libraries/packages associated with the spark pool. - :type custom_libraries: list[~azure.synapse.artifacts.v2020_12_01.models.LibraryInfo] - :param spark_config_properties: Spark configuration file to specify additional properties. - :type spark_config_properties: ~azure.synapse.artifacts.v2020_12_01.models.LibraryRequirements - :param spark_version: The Apache Spark version. - :type spark_version: str - :param default_spark_log_folder: The default folder where Spark logs will be written. - :type default_spark_log_folder: str - :param node_size: The level of compute power that each node in the Big Data pool has. Possible - values include: "None", "Small", "Medium", "Large", "XLarge", "XXLarge", "XXXLarge". - :type node_size: str or ~azure.synapse.artifacts.v2020_12_01.models.NodeSize - :param node_size_family: The kind of nodes that the Big Data pool provides. Possible values - include: "None", "MemoryOptimized". - :type node_size_family: str or ~azure.synapse.artifacts.v2020_12_01.models.NodeSizeFamily - :ivar last_succeeded_timestamp: The time when the Big Data pool was updated successfully. - :vartype last_succeeded_timestamp: ~datetime.datetime - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'location': {'required': True}, - 'last_succeeded_timestamp': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'tags': {'key': 'tags', 'type': '{str}'}, - 'location': {'key': 'location', 'type': 'str'}, - 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, - 'auto_scale': {'key': 'properties.autoScale', 'type': 'AutoScaleProperties'}, - 'creation_date': {'key': 'properties.creationDate', 'type': 'iso-8601'}, - 'auto_pause': {'key': 'properties.autoPause', 'type': 'AutoPauseProperties'}, - 'is_compute_isolation_enabled': {'key': 'properties.isComputeIsolationEnabled', 'type': 'bool'}, - 'session_level_packages_enabled': {'key': 'properties.sessionLevelPackagesEnabled', 'type': 'bool'}, - 'cache_size': {'key': 'properties.cacheSize', 'type': 'int'}, - 'dynamic_executor_allocation': {'key': 'properties.dynamicExecutorAllocation', 'type': 'DynamicExecutorAllocation'}, - 'spark_events_folder': {'key': 'properties.sparkEventsFolder', 'type': 'str'}, - 'node_count': {'key': 'properties.nodeCount', 'type': 'int'}, - 'library_requirements': {'key': 'properties.libraryRequirements', 'type': 'LibraryRequirements'}, - 'custom_libraries': {'key': 'properties.customLibraries', 'type': '[LibraryInfo]'}, - 'spark_config_properties': {'key': 'properties.sparkConfigProperties', 'type': 'LibraryRequirements'}, - 'spark_version': {'key': 'properties.sparkVersion', 'type': 'str'}, - 'default_spark_log_folder': {'key': 'properties.defaultSparkLogFolder', 'type': 'str'}, - 'node_size': {'key': 'properties.nodeSize', 'type': 'str'}, - 'node_size_family': {'key': 'properties.nodeSizeFamily', 'type': 'str'}, - 'last_succeeded_timestamp': {'key': 'properties.lastSucceededTimestamp', 'type': 'iso-8601'}, - } - - def __init__( - self, - *, - location: str, - tags: Optional[Dict[str, str]] = None, - provisioning_state: Optional[str] = None, - auto_scale: Optional["AutoScaleProperties"] = None, - creation_date: Optional[datetime.datetime] = None, - auto_pause: Optional["AutoPauseProperties"] = None, - is_compute_isolation_enabled: Optional[bool] = None, - session_level_packages_enabled: Optional[bool] = None, - cache_size: Optional[int] = None, - dynamic_executor_allocation: Optional["DynamicExecutorAllocation"] = None, - spark_events_folder: Optional[str] = None, - node_count: Optional[int] = None, - library_requirements: Optional["LibraryRequirements"] = None, - custom_libraries: Optional[List["LibraryInfo"]] = None, - spark_config_properties: Optional["LibraryRequirements"] = None, - spark_version: Optional[str] = None, - default_spark_log_folder: Optional[str] = None, - node_size: Optional[Union[str, "NodeSize"]] = None, - node_size_family: Optional[Union[str, "NodeSizeFamily"]] = None, - **kwargs - ): - super(BigDataPoolResourceInfo, self).__init__(tags=tags, location=location, **kwargs) - self.provisioning_state = provisioning_state - self.auto_scale = auto_scale - self.creation_date = creation_date - self.auto_pause = auto_pause - self.is_compute_isolation_enabled = is_compute_isolation_enabled - self.session_level_packages_enabled = session_level_packages_enabled - self.cache_size = cache_size - self.dynamic_executor_allocation = dynamic_executor_allocation - self.spark_events_folder = spark_events_folder - self.node_count = node_count - self.library_requirements = library_requirements - self.custom_libraries = custom_libraries - self.spark_config_properties = spark_config_properties - self.spark_version = spark_version - self.default_spark_log_folder = default_spark_log_folder - self.node_size = node_size - self.node_size_family = node_size_family - self.last_succeeded_timestamp = None - - -class BigDataPoolResourceInfoListResult(msrest.serialization.Model): - """Collection of Big Data pool information. - - :param next_link: Link to the next page of results. - :type next_link: str - :param value: List of Big Data pools. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.BigDataPoolResourceInfo] - """ - - _attribute_map = { - 'next_link': {'key': 'nextLink', 'type': 'str'}, - 'value': {'key': 'value', 'type': '[BigDataPoolResourceInfo]'}, - } - - def __init__( - self, - *, - next_link: Optional[str] = None, - value: Optional[List["BigDataPoolResourceInfo"]] = None, - **kwargs - ): - super(BigDataPoolResourceInfoListResult, self).__init__(**kwargs) - self.next_link = next_link - self.value = value - - -class BinaryDataset(Dataset): - """Binary dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param location: The location of the Binary storage. - :type location: ~azure.synapse.artifacts.v2020_12_01.models.DatasetLocation - :param compression: The data compression method used for the binary dataset. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'location': {'key': 'typeProperties.location', 'type': 'DatasetLocation'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - location: Optional["DatasetLocation"] = None, - compression: Optional["DatasetCompression"] = None, - **kwargs - ): - super(BinaryDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'Binary' # type: str - self.location = location - self.compression = compression - - -class FormatReadSettings(msrest.serialization.Model): - """Format read settings. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: BinaryReadSettings, DelimitedTextReadSettings, JsonReadSettings, XmlReadSettings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - _subtype_map = { - 'type': {'BinaryReadSettings': 'BinaryReadSettings', 'DelimitedTextReadSettings': 'DelimitedTextReadSettings', 'JsonReadSettings': 'JsonReadSettings', 'XmlReadSettings': 'XmlReadSettings'} - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(FormatReadSettings, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = 'FormatReadSettings' # type: str - - -class BinaryReadSettings(FormatReadSettings): - """Binary read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param compression_properties: Compression settings. - :type compression_properties: - ~azure.synapse.artifacts.v2020_12_01.models.CompressionReadSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'compression_properties': {'key': 'compressionProperties', 'type': 'CompressionReadSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - compression_properties: Optional["CompressionReadSettings"] = None, - **kwargs - ): - super(BinaryReadSettings, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'BinaryReadSettings' # type: str - self.compression_properties = compression_properties - - -class BinarySink(CopySink): - """A copy activity Binary sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Binary store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreWriteSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreWriteSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - store_settings: Optional["StoreWriteSettings"] = None, - **kwargs - ): - super(BinarySink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'BinarySink' # type: str - self.store_settings = store_settings - - -class BinarySource(CopySource): - """A copy activity Binary source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Binary store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - :param format_settings: Binary format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.BinaryReadSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'format_settings': {'key': 'formatSettings', 'type': 'BinaryReadSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - store_settings: Optional["StoreReadSettings"] = None, - format_settings: Optional["BinaryReadSettings"] = None, - **kwargs - ): - super(BinarySource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'BinarySource' # type: str - self.store_settings = store_settings - self.format_settings = format_settings - - -class Trigger(msrest.serialization.Model): - """Azure Synapse nested object which contains information about creating pipeline run. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: ChainingTrigger, MultiplePipelineTrigger, RerunTumblingWindowTrigger, TumblingWindowTrigger. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Trigger type.Constant filled by server. - :type type: str - :param description: Trigger description. - :type description: str - :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are - called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRuntimeState - :param annotations: List of tags that can be used for describing the trigger. - :type annotations: list[any] - """ - - _validation = { - 'type': {'required': True}, - 'runtime_state': {'readonly': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - } - - _subtype_map = { - 'type': {'ChainingTrigger': 'ChainingTrigger', 'MultiplePipelineTrigger': 'MultiplePipelineTrigger', 'RerunTumblingWindowTrigger': 'RerunTumblingWindowTrigger', 'TumblingWindowTrigger': 'TumblingWindowTrigger'} - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - annotations: Optional[List[Any]] = None, - **kwargs - ): - super(Trigger, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = 'Trigger' # type: str - self.description = description - self.runtime_state = None - self.annotations = annotations - - -class MultiplePipelineTrigger(Trigger): - """Base class for all triggers that support one to many model for trigger to pipeline. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: BlobEventsTrigger, BlobTrigger, CustomEventsTrigger, ScheduleTrigger. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Trigger type.Constant filled by server. - :type type: str - :param description: Trigger description. - :type description: str - :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are - called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRuntimeState - :param annotations: List of tags that can be used for describing the trigger. - :type annotations: list[any] - :param pipelines: Pipelines that need to be started. - :type pipelines: list[~azure.synapse.artifacts.v2020_12_01.models.TriggerPipelineReference] - """ - - _validation = { - 'type': {'required': True}, - 'runtime_state': {'readonly': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, - } - - _subtype_map = { - 'type': {'BlobEventsTrigger': 'BlobEventsTrigger', 'BlobTrigger': 'BlobTrigger', 'CustomEventsTrigger': 'CustomEventsTrigger', 'ScheduleTrigger': 'ScheduleTrigger'} - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - annotations: Optional[List[Any]] = None, - pipelines: Optional[List["TriggerPipelineReference"]] = None, - **kwargs - ): - super(MultiplePipelineTrigger, self).__init__(additional_properties=additional_properties, description=description, annotations=annotations, **kwargs) - self.type = 'MultiplePipelineTrigger' # type: str - self.pipelines = pipelines - - -class BlobEventsTrigger(MultiplePipelineTrigger): - """Trigger that runs every time a Blob event occurs. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Trigger type.Constant filled by server. - :type type: str - :param description: Trigger description. - :type description: str - :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are - called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRuntimeState - :param annotations: List of tags that can be used for describing the trigger. - :type annotations: list[any] - :param pipelines: Pipelines that need to be started. - :type pipelines: list[~azure.synapse.artifacts.v2020_12_01.models.TriggerPipelineReference] - :param blob_path_begins_with: The blob path must begin with the pattern provided for trigger to - fire. For example, '/records/blobs/december/' will only fire the trigger for blobs in the - december folder under the records container. At least one of these must be provided: - blobPathBeginsWith, blobPathEndsWith. - :type blob_path_begins_with: str - :param blob_path_ends_with: The blob path must end with the pattern provided for trigger to - fire. For example, 'december/boxes.csv' will only fire the trigger for blobs named boxes in a - december folder. At least one of these must be provided: blobPathBeginsWith, blobPathEndsWith. - :type blob_path_ends_with: str - :param ignore_empty_blobs: If set to true, blobs with zero bytes will be ignored. - :type ignore_empty_blobs: bool - :param events: Required. The type of events that cause this trigger to fire. - :type events: list[str or ~azure.synapse.artifacts.v2020_12_01.models.BlobEventType] - :param scope: Required. The ARM resource ID of the Storage Account. - :type scope: str - """ - - _validation = { - 'type': {'required': True}, - 'runtime_state': {'readonly': True}, - 'events': {'required': True}, - 'scope': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, - 'blob_path_begins_with': {'key': 'typeProperties.blobPathBeginsWith', 'type': 'str'}, - 'blob_path_ends_with': {'key': 'typeProperties.blobPathEndsWith', 'type': 'str'}, - 'ignore_empty_blobs': {'key': 'typeProperties.ignoreEmptyBlobs', 'type': 'bool'}, - 'events': {'key': 'typeProperties.events', 'type': '[str]'}, - 'scope': {'key': 'typeProperties.scope', 'type': 'str'}, - } - - def __init__( - self, - *, - events: List[Union[str, "BlobEventType"]], - scope: str, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - annotations: Optional[List[Any]] = None, - pipelines: Optional[List["TriggerPipelineReference"]] = None, - blob_path_begins_with: Optional[str] = None, - blob_path_ends_with: Optional[str] = None, - ignore_empty_blobs: Optional[bool] = None, - **kwargs - ): - super(BlobEventsTrigger, self).__init__(additional_properties=additional_properties, description=description, annotations=annotations, pipelines=pipelines, **kwargs) - self.type = 'BlobEventsTrigger' # type: str - self.blob_path_begins_with = blob_path_begins_with - self.blob_path_ends_with = blob_path_ends_with - self.ignore_empty_blobs = ignore_empty_blobs - self.events = events - self.scope = scope - - -class BlobSink(CopySink): - """A copy activity Azure Blob sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param blob_writer_overwrite_files: Blob writer overwrite files. Type: boolean (or Expression - with resultType boolean). - :type blob_writer_overwrite_files: any - :param blob_writer_date_time_format: Blob writer date time format. Type: string (or Expression - with resultType string). - :type blob_writer_date_time_format: any - :param blob_writer_add_header: Blob writer add header. Type: boolean (or Expression with - resultType boolean). - :type blob_writer_add_header: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'blob_writer_overwrite_files': {'key': 'blobWriterOverwriteFiles', 'type': 'object'}, - 'blob_writer_date_time_format': {'key': 'blobWriterDateTimeFormat', 'type': 'object'}, - 'blob_writer_add_header': {'key': 'blobWriterAddHeader', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - blob_writer_overwrite_files: Optional[Any] = None, - blob_writer_date_time_format: Optional[Any] = None, - blob_writer_add_header: Optional[Any] = None, - copy_behavior: Optional[Any] = None, - **kwargs - ): - super(BlobSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'BlobSink' # type: str - self.blob_writer_overwrite_files = blob_writer_overwrite_files - self.blob_writer_date_time_format = blob_writer_date_time_format - self.blob_writer_add_header = blob_writer_add_header - self.copy_behavior = copy_behavior - - -class BlobSource(CopySource): - """A copy activity Azure Blob source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param treat_empty_as_null: Treat empty as null. Type: boolean (or Expression with resultType - boolean). - :type treat_empty_as_null: any - :param skip_header_line_count: Number of header lines to skip from each blob. Type: integer (or - Expression with resultType integer). - :type skip_header_line_count: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'treat_empty_as_null': {'key': 'treatEmptyAsNull', 'type': 'object'}, - 'skip_header_line_count': {'key': 'skipHeaderLineCount', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - treat_empty_as_null: Optional[Any] = None, - skip_header_line_count: Optional[Any] = None, - recursive: Optional[Any] = None, - **kwargs - ): - super(BlobSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'BlobSource' # type: str - self.treat_empty_as_null = treat_empty_as_null - self.skip_header_line_count = skip_header_line_count - self.recursive = recursive - - -class BlobTrigger(MultiplePipelineTrigger): - """Trigger that runs every time the selected Blob container changes. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Trigger type.Constant filled by server. - :type type: str - :param description: Trigger description. - :type description: str - :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are - called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRuntimeState - :param annotations: List of tags that can be used for describing the trigger. - :type annotations: list[any] - :param pipelines: Pipelines that need to be started. - :type pipelines: list[~azure.synapse.artifacts.v2020_12_01.models.TriggerPipelineReference] - :param folder_path: Required. The path of the container/folder that will trigger the pipeline. - :type folder_path: str - :param max_concurrency: Required. The max number of parallel files to handle when it is - triggered. - :type max_concurrency: int - :param linked_service: Required. The Azure Storage linked service reference. - :type linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - """ - - _validation = { - 'type': {'required': True}, - 'runtime_state': {'readonly': True}, - 'folder_path': {'required': True}, - 'max_concurrency': {'required': True}, - 'linked_service': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, - 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'str'}, - 'max_concurrency': {'key': 'typeProperties.maxConcurrency', 'type': 'int'}, - 'linked_service': {'key': 'typeProperties.linkedService', 'type': 'LinkedServiceReference'}, - } - - def __init__( - self, - *, - folder_path: str, - max_concurrency: int, - linked_service: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - annotations: Optional[List[Any]] = None, - pipelines: Optional[List["TriggerPipelineReference"]] = None, - **kwargs - ): - super(BlobTrigger, self).__init__(additional_properties=additional_properties, description=description, annotations=annotations, pipelines=pipelines, **kwargs) - self.type = 'BlobTrigger' # type: str - self.folder_path = folder_path - self.max_concurrency = max_concurrency - self.linked_service = linked_service - - -class CassandraLinkedService(LinkedService): - """Linked service for Cassandra data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. Host name for connection. Type: string (or Expression with resultType - string). - :type host: any - :param authentication_type: AuthenticationType to be used for connection. Type: string (or - Expression with resultType string). - :type authentication_type: any - :param port: The port for the connection. Type: integer (or Expression with resultType - integer). - :type port: any - :param username: Username for authentication. Type: string (or Expression with resultType - string). - :type username: any - :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - host: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - authentication_type: Optional[Any] = None, - port: Optional[Any] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(CassandraLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Cassandra' # type: str - self.host = host - self.authentication_type = authentication_type - self.port = port - self.username = username - self.password = password - self.encrypted_credential = encrypted_credential - - -class CassandraSource(TabularSource): - """A copy activity source for a Cassandra database. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Should be a SQL-92 query expression or Cassandra Query Language - (CQL) command. Type: string (or Expression with resultType string). - :type query: any - :param consistency_level: The consistency level specifies how many Cassandra servers must - respond to a read request before returning data to the client application. Cassandra checks the - specified number of Cassandra servers for data to satisfy the read request. Must be one of - cassandraSourceReadConsistencyLevels. The default value is 'ONE'. It is case-insensitive. - Possible values include: "ALL", "EACH_QUORUM", "QUORUM", "LOCAL_QUORUM", "ONE", "TWO", "THREE", - "LOCAL_ONE", "SERIAL", "LOCAL_SERIAL". - :type consistency_level: str or - ~azure.synapse.artifacts.v2020_12_01.models.CassandraSourceReadConsistencyLevels - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'consistency_level': {'key': 'consistencyLevel', 'type': 'str'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - consistency_level: Optional[Union[str, "CassandraSourceReadConsistencyLevels"]] = None, - **kwargs - ): - super(CassandraSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'CassandraSource' # type: str - self.query = query - self.consistency_level = consistency_level - - -class CassandraTableDataset(Dataset): - """The Cassandra database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name of the Cassandra database. Type: string (or Expression with - resultType string). - :type table_name: any - :param keyspace: The keyspace of the Cassandra database. Type: string (or Expression with - resultType string). - :type keyspace: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'keyspace': {'key': 'typeProperties.keyspace', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - keyspace: Optional[Any] = None, - **kwargs - ): - super(CassandraTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'CassandraTable' # type: str - self.table_name = table_name - self.keyspace = keyspace - - -class ChainingTrigger(Trigger): - """Trigger that allows the referenced pipeline to depend on other pipeline runs based on runDimension Name/Value pairs. Upstream pipelines should declare the same runDimension Name and their runs should have the values for those runDimensions. The referenced pipeline run would be triggered if the values for the runDimension match for all upstream pipeline runs. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Trigger type.Constant filled by server. - :type type: str - :param description: Trigger description. - :type description: str - :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are - called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRuntimeState - :param annotations: List of tags that can be used for describing the trigger. - :type annotations: list[any] - :param pipeline: Required. Pipeline for which runs are created when all upstream pipelines - complete successfully. - :type pipeline: ~azure.synapse.artifacts.v2020_12_01.models.TriggerPipelineReference - :param depends_on: Required. Upstream Pipelines. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.PipelineReference] - :param run_dimension: Required. Run Dimension property that needs to be emitted by upstream - pipelines. - :type run_dimension: str - """ - - _validation = { - 'type': {'required': True}, - 'runtime_state': {'readonly': True}, - 'pipeline': {'required': True}, - 'depends_on': {'required': True}, - 'run_dimension': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'pipeline': {'key': 'pipeline', 'type': 'TriggerPipelineReference'}, - 'depends_on': {'key': 'typeProperties.dependsOn', 'type': '[PipelineReference]'}, - 'run_dimension': {'key': 'typeProperties.runDimension', 'type': 'str'}, - } - - def __init__( - self, - *, - pipeline: "TriggerPipelineReference", - depends_on: List["PipelineReference"], - run_dimension: str, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - annotations: Optional[List[Any]] = None, - **kwargs - ): - super(ChainingTrigger, self).__init__(additional_properties=additional_properties, description=description, annotations=annotations, **kwargs) - self.type = 'ChainingTrigger' # type: str - self.pipeline = pipeline - self.depends_on = depends_on - self.run_dimension = run_dimension - - -class CloudError(msrest.serialization.Model): - """The object that defines the structure of an Azure Synapse error response. - - All required parameters must be populated in order to send to Azure. - - :param code: Required. Error code. - :type code: str - :param message: Required. Error message. - :type message: str - :param target: Property name/path in request associated with error. - :type target: str - :param details: Array with additional error details. - :type details: list[~azure.synapse.artifacts.v2020_12_01.models.CloudError] - """ - - _validation = { - 'code': {'required': True}, - 'message': {'required': True}, - } - - _attribute_map = { - 'code': {'key': 'error.code', 'type': 'str'}, - 'message': {'key': 'error.message', 'type': 'str'}, - 'target': {'key': 'error.target', 'type': 'str'}, - 'details': {'key': 'error.details', 'type': '[CloudError]'}, - } - - def __init__( - self, - *, - code: str, - message: str, - target: Optional[str] = None, - details: Optional[List["CloudError"]] = None, - **kwargs - ): - super(CloudError, self).__init__(**kwargs) - self.code = code - self.message = message - self.target = target - self.details = details - - -class CommonDataServiceForAppsEntityDataset(Dataset): - """The Common Data Service for Apps entity dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param entity_name: The logical name of the entity. Type: string (or Expression with resultType - string). - :type entity_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'entity_name': {'key': 'typeProperties.entityName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - entity_name: Optional[Any] = None, - **kwargs - ): - super(CommonDataServiceForAppsEntityDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'CommonDataServiceForAppsEntity' # type: str - self.entity_name = entity_name - - -class CommonDataServiceForAppsLinkedService(LinkedService): - """Common Data Service for Apps linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param deployment_type: Required. The deployment type of the Common Data Service for Apps - instance. 'Online' for Common Data Service for Apps Online and 'OnPremisesWithIfd' for Common - Data Service for Apps on-premises with Ifd. Type: string (or Expression with resultType - string). Possible values include: "Online", "OnPremisesWithIfd". - :type deployment_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsDeploymentType - :param host_name: The host name of the on-premises Common Data Service for Apps server. The - property is required for on-prem and not allowed for online. Type: string (or Expression with - resultType string). - :type host_name: any - :param port: The port of on-premises Common Data Service for Apps server. The property is - required for on-prem and not allowed for online. Default is 443. Type: integer (or Expression - with resultType integer), minimum: 0. - :type port: any - :param service_uri: The URL to the Microsoft Common Data Service for Apps server. The property - is required for on-line and not allowed for on-prem. Type: string (or Expression with - resultType string). - :type service_uri: any - :param organization_name: The organization name of the Common Data Service for Apps instance. - The property is required for on-prem and required for online when there are more than one - Common Data Service for Apps instances associated with the user. Type: string (or Expression - with resultType string). - :type organization_name: any - :param authentication_type: Required. The authentication type to connect to Common Data Service - for Apps server. 'Office365' for online scenario, 'Ifd' for on-premises with Ifd scenario. - 'AADServicePrincipal' for Server-To-Server authentication in online scenario. Type: string (or - Expression with resultType string). Possible values include: "Office365", "Ifd", - "AADServicePrincipal". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsAuthenticationType - :param username: User name to access the Common Data Service for Apps instance. Type: string - (or Expression with resultType string). - :type username: any - :param password: Password to access the Common Data Service for Apps instance. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param service_principal_id: The client ID of the application in Azure Active Directory used - for Server-To-Server authentication. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_credential_type: A string from ServicePrincipalCredentialEnum or an - expression. - :type service_principal_credential_type: any - :param service_principal_credential: The credential of the service principal object in Azure - Active Directory. If servicePrincipalCredentialType is 'ServicePrincipalKey', - servicePrincipalCredential can be SecureString or AzureKeyVaultSecretReference. If - servicePrincipalCredentialType is 'ServicePrincipalCert', servicePrincipalCredential can only - be AzureKeyVaultSecretReference. - :type service_principal_credential: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'deployment_type': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'str'}, - 'host_name': {'key': 'typeProperties.hostName', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'service_uri': {'key': 'typeProperties.serviceUri', 'type': 'object'}, - 'organization_name': {'key': 'typeProperties.organizationName', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_credential_type': {'key': 'typeProperties.servicePrincipalCredentialType', 'type': 'object'}, - 'service_principal_credential': {'key': 'typeProperties.servicePrincipalCredential', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - deployment_type: Union[str, "DynamicsDeploymentType"], - authentication_type: Union[str, "DynamicsAuthenticationType"], - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - host_name: Optional[Any] = None, - port: Optional[Any] = None, - service_uri: Optional[Any] = None, - organization_name: Optional[Any] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - service_principal_id: Optional[Any] = None, - service_principal_credential_type: Optional[Any] = None, - service_principal_credential: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(CommonDataServiceForAppsLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'CommonDataServiceForApps' # type: str - self.deployment_type = deployment_type - self.host_name = host_name - self.port = port - self.service_uri = service_uri - self.organization_name = organization_name - self.authentication_type = authentication_type - self.username = username - self.password = password - self.service_principal_id = service_principal_id - self.service_principal_credential_type = service_principal_credential_type - self.service_principal_credential = service_principal_credential - self.encrypted_credential = encrypted_credential - - -class CommonDataServiceForAppsSink(CopySink): - """A copy activity Common Data Service for Apps sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param write_behavior: Required. The write behavior for the operation. Possible values include: - "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsSinkWriteBehavior - :param ignore_null_values: The flag indicating whether to ignore null values from input dataset - (except key fields) during write operation. Default is false. Type: boolean (or Expression with - resultType boolean). - :type ignore_null_values: any - :param alternate_key_name: The logical name of the alternate key which will be used when - upserting records. Type: string (or Expression with resultType string). - :type alternate_key_name: any - """ - - _validation = { - 'type': {'required': True}, - 'write_behavior': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, - 'ignore_null_values': {'key': 'ignoreNullValues', 'type': 'object'}, - 'alternate_key_name': {'key': 'alternateKeyName', 'type': 'object'}, - } - - def __init__( - self, - *, - write_behavior: Union[str, "DynamicsSinkWriteBehavior"], - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - ignore_null_values: Optional[Any] = None, - alternate_key_name: Optional[Any] = None, - **kwargs - ): - super(CommonDataServiceForAppsSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'CommonDataServiceForAppsSink' # type: str - self.write_behavior = write_behavior - self.ignore_null_values = ignore_null_values - self.alternate_key_name = alternate_key_name - - -class CommonDataServiceForAppsSource(CopySource): - """A copy activity Common Data Service for Apps source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: FetchXML is a proprietary query language that is used in Microsoft Common Data - Service for Apps (online & on-premises). Type: string (or Expression with resultType string). - :type query: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query: Optional[Any] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(CommonDataServiceForAppsSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'CommonDataServiceForAppsSource' # type: str - self.query = query - self.additional_columns = additional_columns - - -class CompressionReadSettings(msrest.serialization.Model): - """Compression read settings. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: TarGZipReadSettings, TarReadSettings, ZipDeflateReadSettings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The Compression setting type.Constant filled by server. - :type type: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - _subtype_map = { - 'type': {'TarGZipReadSettings': 'TarGZipReadSettings', 'TarReadSettings': 'TarReadSettings', 'ZipDeflateReadSettings': 'ZipDeflateReadSettings'} - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(CompressionReadSettings, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = 'CompressionReadSettings' # type: str - - -class ConcurLinkedService(LinkedService): - """Concur Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_properties: Properties used to connect to Concur. It is mutually exclusive - with any other properties in the linked service. Type: object. - :type connection_properties: any - :param client_id: Required. Application client_id supplied by Concur App Management. - :type client_id: any - :param username: Required. The user name that you use to access Concur Service. - :type username: any - :param password: The password corresponding to the user name that you provided in the username - field. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'client_id': {'required': True}, - 'username': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_properties': {'key': 'typeProperties.connectionProperties', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - client_id: Any, - username: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_properties: Optional[Any] = None, - password: Optional["SecretBase"] = None, - use_encrypted_endpoints: Optional[Any] = None, - use_host_verification: Optional[Any] = None, - use_peer_verification: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(ConcurLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Concur' # type: str - self.connection_properties = connection_properties - self.client_id = client_id - self.username = username - self.password = password - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential - - -class ConcurObjectDataset(Dataset): - """Concur Service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(ConcurObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'ConcurObject' # type: str - self.table_name = table_name - - -class ConcurSource(TabularSource): - """A copy activity Concur Service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(ConcurSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'ConcurSource' # type: str - self.query = query - - -class CopyActivity(ExecutionActivity): - """Copy activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param inputs: List of inputs for the activity. - :type inputs: list[~azure.synapse.artifacts.v2020_12_01.models.DatasetReference] - :param outputs: List of outputs for the activity. - :type outputs: list[~azure.synapse.artifacts.v2020_12_01.models.DatasetReference] - :param source: Required. Copy activity source. - :type source: ~azure.synapse.artifacts.v2020_12_01.models.CopySource - :param sink: Required. Copy activity sink. - :type sink: ~azure.synapse.artifacts.v2020_12_01.models.CopySink - :param translator: Copy activity translator. If not specified, tabular translator is used. - :type translator: any - :param enable_staging: Specifies whether to copy data via an interim staging. Default value is - false. Type: boolean (or Expression with resultType boolean). - :type enable_staging: any - :param staging_settings: Specifies interim staging settings when EnableStaging is true. - :type staging_settings: ~azure.synapse.artifacts.v2020_12_01.models.StagingSettings - :param parallel_copies: Maximum number of concurrent sessions opened on the source or sink to - avoid overloading the data store. Type: integer (or Expression with resultType integer), - minimum: 0. - :type parallel_copies: any - :param data_integration_units: Maximum number of data integration units that can be used to - perform this data movement. Type: integer (or Expression with resultType integer), minimum: 0. - :type data_integration_units: any - :param enable_skip_incompatible_row: Whether to skip incompatible row. Default value is false. - Type: boolean (or Expression with resultType boolean). - :type enable_skip_incompatible_row: any - :param redirect_incompatible_row_settings: Redirect incompatible row settings when - EnableSkipIncompatibleRow is true. - :type redirect_incompatible_row_settings: - ~azure.synapse.artifacts.v2020_12_01.models.RedirectIncompatibleRowSettings - :param log_storage_settings: (Deprecated. Please use LogSettings) Log storage settings customer - need to provide when enabling session log. - :type log_storage_settings: ~azure.synapse.artifacts.v2020_12_01.models.LogStorageSettings - :param log_settings: Log settings customer needs provide when enabling log. - :type log_settings: ~azure.synapse.artifacts.v2020_12_01.models.LogSettings - :param preserve_rules: Preserve Rules. - :type preserve_rules: list[any] - :param preserve: Preserve rules. - :type preserve: list[any] - :param validate_data_consistency: Whether to enable Data Consistency validation. Type: boolean - (or Expression with resultType boolean). - :type validate_data_consistency: any - :param skip_error_file: Specify the fault tolerance for data consistency. - :type skip_error_file: ~azure.synapse.artifacts.v2020_12_01.models.SkipErrorFile - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'source': {'required': True}, - 'sink': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'inputs': {'key': 'inputs', 'type': '[DatasetReference]'}, - 'outputs': {'key': 'outputs', 'type': '[DatasetReference]'}, - 'source': {'key': 'typeProperties.source', 'type': 'CopySource'}, - 'sink': {'key': 'typeProperties.sink', 'type': 'CopySink'}, - 'translator': {'key': 'typeProperties.translator', 'type': 'object'}, - 'enable_staging': {'key': 'typeProperties.enableStaging', 'type': 'object'}, - 'staging_settings': {'key': 'typeProperties.stagingSettings', 'type': 'StagingSettings'}, - 'parallel_copies': {'key': 'typeProperties.parallelCopies', 'type': 'object'}, - 'data_integration_units': {'key': 'typeProperties.dataIntegrationUnits', 'type': 'object'}, - 'enable_skip_incompatible_row': {'key': 'typeProperties.enableSkipIncompatibleRow', 'type': 'object'}, - 'redirect_incompatible_row_settings': {'key': 'typeProperties.redirectIncompatibleRowSettings', 'type': 'RedirectIncompatibleRowSettings'}, - 'log_storage_settings': {'key': 'typeProperties.logStorageSettings', 'type': 'LogStorageSettings'}, - 'log_settings': {'key': 'typeProperties.logSettings', 'type': 'LogSettings'}, - 'preserve_rules': {'key': 'typeProperties.preserveRules', 'type': '[object]'}, - 'preserve': {'key': 'typeProperties.preserve', 'type': '[object]'}, - 'validate_data_consistency': {'key': 'typeProperties.validateDataConsistency', 'type': 'object'}, - 'skip_error_file': {'key': 'typeProperties.skipErrorFile', 'type': 'SkipErrorFile'}, - } - - def __init__( - self, - *, - name: str, - source: "CopySource", - sink: "CopySink", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - inputs: Optional[List["DatasetReference"]] = None, - outputs: Optional[List["DatasetReference"]] = None, - translator: Optional[Any] = None, - enable_staging: Optional[Any] = None, - staging_settings: Optional["StagingSettings"] = None, - parallel_copies: Optional[Any] = None, - data_integration_units: Optional[Any] = None, - enable_skip_incompatible_row: Optional[Any] = None, - redirect_incompatible_row_settings: Optional["RedirectIncompatibleRowSettings"] = None, - log_storage_settings: Optional["LogStorageSettings"] = None, - log_settings: Optional["LogSettings"] = None, - preserve_rules: Optional[List[Any]] = None, - preserve: Optional[List[Any]] = None, - validate_data_consistency: Optional[Any] = None, - skip_error_file: Optional["SkipErrorFile"] = None, - **kwargs - ): - super(CopyActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'Copy' # type: str - self.inputs = inputs - self.outputs = outputs - self.source = source - self.sink = sink - self.translator = translator - self.enable_staging = enable_staging - self.staging_settings = staging_settings - self.parallel_copies = parallel_copies - self.data_integration_units = data_integration_units - self.enable_skip_incompatible_row = enable_skip_incompatible_row - self.redirect_incompatible_row_settings = redirect_incompatible_row_settings - self.log_storage_settings = log_storage_settings - self.log_settings = log_settings - self.preserve_rules = preserve_rules - self.preserve = preserve - self.validate_data_consistency = validate_data_consistency - self.skip_error_file = skip_error_file - - -class CopyActivityLogSettings(msrest.serialization.Model): - """Settings for copy activity log. - - :param log_level: Gets or sets the log level, support: Info, Warning. Type: string (or - Expression with resultType string). - :type log_level: any - :param enable_reliable_logging: Specifies whether to enable reliable logging. Type: boolean (or - Expression with resultType boolean). - :type enable_reliable_logging: any - """ - - _attribute_map = { - 'log_level': {'key': 'logLevel', 'type': 'object'}, - 'enable_reliable_logging': {'key': 'enableReliableLogging', 'type': 'object'}, - } - - def __init__( - self, - *, - log_level: Optional[Any] = None, - enable_reliable_logging: Optional[Any] = None, - **kwargs - ): - super(CopyActivityLogSettings, self).__init__(**kwargs) - self.log_level = log_level - self.enable_reliable_logging = enable_reliable_logging - - -class CopyTranslator(msrest.serialization.Model): - """A copy activity translator. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: TabularTranslator. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy translator type.Constant filled by server. - :type type: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - _subtype_map = { - 'type': {'TabularTranslator': 'TabularTranslator'} - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(CopyTranslator, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = 'CopyTranslator' # type: str - - -class CosmosDbLinkedService(LinkedService): - """Microsoft Azure Cosmos Database (CosmosDB) linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: The connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param account_endpoint: The endpoint of the Azure CosmosDB account. Type: string (or - Expression with resultType string). - :type account_endpoint: any - :param database: The name of the database. Type: string (or Expression with resultType string). - :type database: any - :param account_key: The account key of the Azure CosmosDB account. Type: SecureString or - AzureKeyVaultSecretReference. - :type account_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'account_endpoint': {'key': 'typeProperties.accountEndpoint', 'type': 'object'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'account_key': {'key': 'typeProperties.accountKey', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_string: Optional[Any] = None, - account_endpoint: Optional[Any] = None, - database: Optional[Any] = None, - account_key: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(CosmosDbLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'CosmosDb' # type: str - self.connection_string = connection_string - self.account_endpoint = account_endpoint - self.database = database - self.account_key = account_key - self.encrypted_credential = encrypted_credential - - -class CosmosDbMongoDbApiCollectionDataset(Dataset): - """The CosmosDB (MongoDB API) database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param collection: Required. The collection name of the CosmosDB (MongoDB API) database. Type: - string (or Expression with resultType string). - :type collection: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'collection': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'collection': {'key': 'typeProperties.collection', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - collection: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - **kwargs - ): - super(CosmosDbMongoDbApiCollectionDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'CosmosDbMongoDbApiCollection' # type: str - self.collection = collection - - -class CosmosDbMongoDbApiLinkedService(LinkedService): - """Linked service for CosmosDB (MongoDB API) data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The CosmosDB (MongoDB API) connection string. Type: string, - SecureString or AzureKeyVaultSecretReference. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param database: Required. The name of the CosmosDB (MongoDB API) database that you want to - access. Type: string (or Expression with resultType string). - :type database: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - 'database': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - } - - def __init__( - self, - *, - connection_string: Any, - database: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - **kwargs - ): - super(CosmosDbMongoDbApiLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'CosmosDbMongoDbApi' # type: str - self.connection_string = connection_string - self.database = database - - -class CosmosDbMongoDbApiSink(CopySink): - """A copy activity sink for a CosmosDB (MongoDB API) database. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param write_behavior: Specifies whether the document with same key to be overwritten (upsert) - rather than throw exception (insert). The default value is "insert". Type: string (or - Expression with resultType string). Type: string (or Expression with resultType string). - :type write_behavior: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - write_behavior: Optional[Any] = None, - **kwargs - ): - super(CosmosDbMongoDbApiSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'CosmosDbMongoDbApiSink' # type: str - self.write_behavior = write_behavior - - -class CosmosDbMongoDbApiSource(CopySource): - """A copy activity source for a CosmosDB (MongoDB API) database. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param filter: Specifies selection filter using query operators. To return all documents in a - collection, omit this parameter or pass an empty document ({}). Type: string (or Expression - with resultType string). - :type filter: any - :param cursor_methods: Cursor methods for Mongodb query. - :type cursor_methods: - ~azure.synapse.artifacts.v2020_12_01.models.MongoDbCursorMethodsProperties - :param batch_size: Specifies the number of documents to return in each batch of the response - from MongoDB instance. In most cases, modifying the batch size will not affect the user or the - application. This property's main purpose is to avoid hit the limitation of response size. - Type: integer (or Expression with resultType integer). - :type batch_size: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'filter': {'key': 'filter', 'type': 'object'}, - 'cursor_methods': {'key': 'cursorMethods', 'type': 'MongoDbCursorMethodsProperties'}, - 'batch_size': {'key': 'batchSize', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - filter: Optional[Any] = None, - cursor_methods: Optional["MongoDbCursorMethodsProperties"] = None, - batch_size: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(CosmosDbMongoDbApiSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'CosmosDbMongoDbApiSource' # type: str - self.filter = filter - self.cursor_methods = cursor_methods - self.batch_size = batch_size - self.query_timeout = query_timeout - self.additional_columns = additional_columns - - -class CosmosDbSqlApiCollectionDataset(Dataset): - """Microsoft Azure CosmosDB (SQL API) Collection dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param collection_name: Required. CosmosDB (SQL API) collection name. Type: string (or - Expression with resultType string). - :type collection_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'collection_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'collection_name': {'key': 'typeProperties.collectionName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - collection_name: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - **kwargs - ): - super(CosmosDbSqlApiCollectionDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'CosmosDbSqlApiCollection' # type: str - self.collection_name = collection_name - - -class CosmosDbSqlApiSink(CopySink): - """A copy activity Azure CosmosDB (SQL API) Collection sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param write_behavior: Describes how to write data to Azure Cosmos DB. Type: string (or - Expression with resultType string). Allowed values: insert and upsert. - :type write_behavior: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - write_behavior: Optional[Any] = None, - **kwargs - ): - super(CosmosDbSqlApiSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'CosmosDbSqlApiSink' # type: str - self.write_behavior = write_behavior - - -class CosmosDbSqlApiSource(CopySource): - """A copy activity Azure CosmosDB (SQL API) Collection source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: SQL API query. Type: string (or Expression with resultType string). - :type query: any - :param page_size: Page size of the result. Type: integer (or Expression with resultType - integer). - :type page_size: any - :param preferred_regions: Preferred regions. Type: array of strings (or Expression with - resultType array of strings). - :type preferred_regions: any - :param detect_datetime: Whether detect primitive values as datetime values. Type: boolean (or - Expression with resultType boolean). - :type detect_datetime: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'page_size': {'key': 'pageSize', 'type': 'object'}, - 'preferred_regions': {'key': 'preferredRegions', 'type': 'object'}, - 'detect_datetime': {'key': 'detectDatetime', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query: Optional[Any] = None, - page_size: Optional[Any] = None, - preferred_regions: Optional[Any] = None, - detect_datetime: Optional[Any] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(CosmosDbSqlApiSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'CosmosDbSqlApiSource' # type: str - self.query = query - self.page_size = page_size - self.preferred_regions = preferred_regions - self.detect_datetime = detect_datetime - self.additional_columns = additional_columns - - -class CouchbaseLinkedService(LinkedService): - """Couchbase server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: An ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param cred_string: The Azure key vault secret reference of credString in connection string. - :type cred_string: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'cred_string': {'key': 'typeProperties.credString', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_string: Optional[Any] = None, - cred_string: Optional["AzureKeyVaultSecretReference"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(CouchbaseLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Couchbase' # type: str - self.connection_string = connection_string - self.cred_string = cred_string - self.encrypted_credential = encrypted_credential - - -class CouchbaseSource(TabularSource): - """A copy activity Couchbase server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(CouchbaseSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'CouchbaseSource' # type: str - self.query = query - - -class CouchbaseTableDataset(Dataset): - """Couchbase server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(CouchbaseTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'CouchbaseTable' # type: str - self.table_name = table_name - - -class CreateDataFlowDebugSessionRequest(msrest.serialization.Model): - """Request body structure for creating data flow debug session. - - :param data_flow_name: The name of the data flow. - :type data_flow_name: str - :param existing_cluster_id: The ID of existing Databricks cluster. - :type existing_cluster_id: str - :param cluster_timeout: Timeout setting for Databricks cluster. - :type cluster_timeout: int - :param new_cluster_name: The name of new Databricks cluster. - :type new_cluster_name: str - :param new_cluster_node_type: The type of new Databricks cluster. - :type new_cluster_node_type: str - :param data_bricks_linked_service: Data bricks linked service. - :type data_bricks_linked_service: - ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceResource - """ - - _attribute_map = { - 'data_flow_name': {'key': 'dataFlowName', 'type': 'str'}, - 'existing_cluster_id': {'key': 'existingClusterId', 'type': 'str'}, - 'cluster_timeout': {'key': 'clusterTimeout', 'type': 'int'}, - 'new_cluster_name': {'key': 'newClusterName', 'type': 'str'}, - 'new_cluster_node_type': {'key': 'newClusterNodeType', 'type': 'str'}, - 'data_bricks_linked_service': {'key': 'dataBricksLinkedService', 'type': 'LinkedServiceResource'}, - } - - def __init__( - self, - *, - data_flow_name: Optional[str] = None, - existing_cluster_id: Optional[str] = None, - cluster_timeout: Optional[int] = None, - new_cluster_name: Optional[str] = None, - new_cluster_node_type: Optional[str] = None, - data_bricks_linked_service: Optional["LinkedServiceResource"] = None, - **kwargs - ): - super(CreateDataFlowDebugSessionRequest, self).__init__(**kwargs) - self.data_flow_name = data_flow_name - self.existing_cluster_id = existing_cluster_id - self.cluster_timeout = cluster_timeout - self.new_cluster_name = new_cluster_name - self.new_cluster_node_type = new_cluster_node_type - self.data_bricks_linked_service = data_bricks_linked_service - - -class CreateDataFlowDebugSessionResponse(msrest.serialization.Model): - """Response body structure for creating data flow debug session. - - :param session_id: The ID of data flow debug session. - :type session_id: str - """ - - _attribute_map = { - 'session_id': {'key': 'sessionId', 'type': 'str'}, - } - - def __init__( - self, - *, - session_id: Optional[str] = None, - **kwargs - ): - super(CreateDataFlowDebugSessionResponse, self).__init__(**kwargs) - self.session_id = session_id - - -class CreateRunResponse(msrest.serialization.Model): - """Response body with a run identifier. - - All required parameters must be populated in order to send to Azure. - - :param run_id: Required. Identifier of a run. - :type run_id: str - """ - - _validation = { - 'run_id': {'required': True}, - } - - _attribute_map = { - 'run_id': {'key': 'runId', 'type': 'str'}, - } - - def __init__( - self, - *, - run_id: str, - **kwargs - ): - super(CreateRunResponse, self).__init__(**kwargs) - self.run_id = run_id - - -class CustomActivity(ExecutionActivity): - """Custom activity type. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param command: Required. Command for custom activity Type: string (or Expression with - resultType string). - :type command: any - :param resource_linked_service: Resource linked service reference. - :type resource_linked_service: - ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param folder_path: Folder path for resource files Type: string (or Expression with resultType - string). - :type folder_path: any - :param reference_objects: Reference objects. - :type reference_objects: - ~azure.synapse.artifacts.v2020_12_01.models.CustomActivityReferenceObject - :param extended_properties: User defined property bag. There is no restriction on the keys or - values that can be used. The user specified custom activity has the full responsibility to - consume and interpret the content defined. - :type extended_properties: dict[str, any] - :param retention_time_in_days: The retention time for the files submitted for custom activity. - Type: double (or Expression with resultType double). - :type retention_time_in_days: any - :param auto_user_specification: Elevation level and scope for the user, default is nonadmin - task. Type: string (or Expression with resultType double). - :type auto_user_specification: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'command': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'command': {'key': 'typeProperties.command', 'type': 'object'}, - 'resource_linked_service': {'key': 'typeProperties.resourceLinkedService', 'type': 'LinkedServiceReference'}, - 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, - 'reference_objects': {'key': 'typeProperties.referenceObjects', 'type': 'CustomActivityReferenceObject'}, - 'extended_properties': {'key': 'typeProperties.extendedProperties', 'type': '{object}'}, - 'retention_time_in_days': {'key': 'typeProperties.retentionTimeInDays', 'type': 'object'}, - 'auto_user_specification': {'key': 'typeProperties.autoUserSpecification', 'type': 'object'}, - } - - def __init__( - self, - *, - name: str, - command: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - resource_linked_service: Optional["LinkedServiceReference"] = None, - folder_path: Optional[Any] = None, - reference_objects: Optional["CustomActivityReferenceObject"] = None, - extended_properties: Optional[Dict[str, Any]] = None, - retention_time_in_days: Optional[Any] = None, - auto_user_specification: Optional[Any] = None, - **kwargs - ): - super(CustomActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'Custom' # type: str - self.command = command - self.resource_linked_service = resource_linked_service - self.folder_path = folder_path - self.reference_objects = reference_objects - self.extended_properties = extended_properties - self.retention_time_in_days = retention_time_in_days - self.auto_user_specification = auto_user_specification - - -class CustomActivityReferenceObject(msrest.serialization.Model): - """Reference objects for custom activity. - - :param linked_services: Linked service references. - :type linked_services: list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference] - :param datasets: Dataset references. - :type datasets: list[~azure.synapse.artifacts.v2020_12_01.models.DatasetReference] - """ - - _attribute_map = { - 'linked_services': {'key': 'linkedServices', 'type': '[LinkedServiceReference]'}, - 'datasets': {'key': 'datasets', 'type': '[DatasetReference]'}, - } - - def __init__( - self, - *, - linked_services: Optional[List["LinkedServiceReference"]] = None, - datasets: Optional[List["DatasetReference"]] = None, - **kwargs - ): - super(CustomActivityReferenceObject, self).__init__(**kwargs) - self.linked_services = linked_services - self.datasets = datasets - - -class CustomDataset(Dataset): - """The custom dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param type_properties: Custom dataset properties. - :type type_properties: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'type_properties': {'key': 'typeProperties', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - type_properties: Optional[Any] = None, - **kwargs - ): - super(CustomDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'CustomDataset' # type: str - self.type_properties = type_properties - - -class CustomDataSourceLinkedService(LinkedService): - """Custom linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param type_properties: Required. Custom linked service properties. - :type type_properties: any - """ - - _validation = { - 'type': {'required': True}, - 'type_properties': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'type_properties': {'key': 'typeProperties', 'type': 'object'}, - } - - def __init__( - self, - *, - type_properties: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - **kwargs - ): - super(CustomDataSourceLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'CustomDataSource' # type: str - self.type_properties = type_properties - - -class CustomerManagedKeyDetails(msrest.serialization.Model): - """Details of the customer managed key associated with the workspace. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar status: The customer managed key status on the workspace. - :vartype status: str - :param key: The key object of the workspace. - :type key: ~azure.synapse.artifacts.v2020_12_01.models.WorkspaceKeyDetails - """ - - _validation = { - 'status': {'readonly': True}, - } - - _attribute_map = { - 'status': {'key': 'status', 'type': 'str'}, - 'key': {'key': 'key', 'type': 'WorkspaceKeyDetails'}, - } - - def __init__( - self, - *, - key: Optional["WorkspaceKeyDetails"] = None, - **kwargs - ): - super(CustomerManagedKeyDetails, self).__init__(**kwargs) - self.status = None - self.key = key - - -class CustomEventsTrigger(MultiplePipelineTrigger): - """Trigger that runs every time a custom event is received. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Trigger type.Constant filled by server. - :type type: str - :param description: Trigger description. - :type description: str - :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are - called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRuntimeState - :param annotations: List of tags that can be used for describing the trigger. - :type annotations: list[any] - :param pipelines: Pipelines that need to be started. - :type pipelines: list[~azure.synapse.artifacts.v2020_12_01.models.TriggerPipelineReference] - :param subject_begins_with: The event subject must begin with the pattern provided for trigger - to fire. At least one of these must be provided: subjectBeginsWith, subjectEndsWith. - :type subject_begins_with: str - :param subject_ends_with: The event subject must end with the pattern provided for trigger to - fire. At least one of these must be provided: subjectBeginsWith, subjectEndsWith. - :type subject_ends_with: str - :param events: Required. The list of event types that cause this trigger to fire. - :type events: list[any] - :param scope: Required. The ARM resource ID of the Azure Event Grid Topic. - :type scope: str - """ - - _validation = { - 'type': {'required': True}, - 'runtime_state': {'readonly': True}, - 'events': {'required': True}, - 'scope': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, - 'subject_begins_with': {'key': 'typeProperties.subjectBeginsWith', 'type': 'str'}, - 'subject_ends_with': {'key': 'typeProperties.subjectEndsWith', 'type': 'str'}, - 'events': {'key': 'typeProperties.events', 'type': '[object]'}, - 'scope': {'key': 'typeProperties.scope', 'type': 'str'}, - } - - def __init__( - self, - *, - events: List[Any], - scope: str, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - annotations: Optional[List[Any]] = None, - pipelines: Optional[List["TriggerPipelineReference"]] = None, - subject_begins_with: Optional[str] = None, - subject_ends_with: Optional[str] = None, - **kwargs - ): - super(CustomEventsTrigger, self).__init__(additional_properties=additional_properties, description=description, annotations=annotations, pipelines=pipelines, **kwargs) - self.type = 'CustomEventsTrigger' # type: str - self.subject_begins_with = subject_begins_with - self.subject_ends_with = subject_ends_with - self.events = events - self.scope = scope - - -class CustomSetupBase(msrest.serialization.Model): - """The base definition of the custom setup. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: . - - All required parameters must be populated in order to send to Azure. - - :param type: Required. The type of custom setup.Constant filled by server. - :type type: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - } - - _subtype_map = { - 'type': {} - } - - def __init__( - self, - **kwargs - ): - super(CustomSetupBase, self).__init__(**kwargs) - self.type = None # type: Optional[str] - - -class DatabricksNotebookActivity(ExecutionActivity): - """DatabricksNotebook activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param notebook_path: Required. The absolute path of the notebook to be run in the Databricks - Workspace. This path must begin with a slash. Type: string (or Expression with resultType - string). - :type notebook_path: any - :param base_parameters: Base parameters to be used for each run of this job.If the notebook - takes a parameter that is not specified, the default value from the notebook will be used. - :type base_parameters: dict[str, any] - :param libraries: A list of libraries to be installed on the cluster that will execute the job. - :type libraries: list[dict[str, any]] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'notebook_path': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'notebook_path': {'key': 'typeProperties.notebookPath', 'type': 'object'}, - 'base_parameters': {'key': 'typeProperties.baseParameters', 'type': '{object}'}, - 'libraries': {'key': 'typeProperties.libraries', 'type': '[{object}]'}, - } - - def __init__( - self, - *, - name: str, - notebook_path: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - base_parameters: Optional[Dict[str, Any]] = None, - libraries: Optional[List[Dict[str, Any]]] = None, - **kwargs - ): - super(DatabricksNotebookActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'DatabricksNotebook' # type: str - self.notebook_path = notebook_path - self.base_parameters = base_parameters - self.libraries = libraries - - -class DatabricksSparkJarActivity(ExecutionActivity): - """DatabricksSparkJar activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param main_class_name: Required. The full name of the class containing the main method to be - executed. This class must be contained in a JAR provided as a library. Type: string (or - Expression with resultType string). - :type main_class_name: any - :param parameters: Parameters that will be passed to the main method. - :type parameters: list[any] - :param libraries: A list of libraries to be installed on the cluster that will execute the job. - :type libraries: list[dict[str, any]] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'main_class_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'main_class_name': {'key': 'typeProperties.mainClassName', 'type': 'object'}, - 'parameters': {'key': 'typeProperties.parameters', 'type': '[object]'}, - 'libraries': {'key': 'typeProperties.libraries', 'type': '[{object}]'}, - } - - def __init__( - self, - *, - name: str, - main_class_name: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - parameters: Optional[List[Any]] = None, - libraries: Optional[List[Dict[str, Any]]] = None, - **kwargs - ): - super(DatabricksSparkJarActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'DatabricksSparkJar' # type: str - self.main_class_name = main_class_name - self.parameters = parameters - self.libraries = libraries - - -class DatabricksSparkPythonActivity(ExecutionActivity): - """DatabricksSparkPython activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param python_file: Required. The URI of the Python file to be executed. DBFS paths are - supported. Type: string (or Expression with resultType string). - :type python_file: any - :param parameters: Command line parameters that will be passed to the Python file. - :type parameters: list[any] - :param libraries: A list of libraries to be installed on the cluster that will execute the job. - :type libraries: list[dict[str, any]] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'python_file': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'python_file': {'key': 'typeProperties.pythonFile', 'type': 'object'}, - 'parameters': {'key': 'typeProperties.parameters', 'type': '[object]'}, - 'libraries': {'key': 'typeProperties.libraries', 'type': '[{object}]'}, - } - - def __init__( - self, - *, - name: str, - python_file: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - parameters: Optional[List[Any]] = None, - libraries: Optional[List[Dict[str, Any]]] = None, - **kwargs - ): - super(DatabricksSparkPythonActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'DatabricksSparkPython' # type: str - self.python_file = python_file - self.parameters = parameters - self.libraries = libraries - - -class DataFlow(msrest.serialization.Model): - """Azure Synapse nested object which contains a flow with data movements and transformations. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: MappingDataFlow. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Type of data flow.Constant filled by server. - :type type: str - :param description: The description of the data flow. - :type description: str - :param annotations: List of tags that can be used for describing the data flow. - :type annotations: list[any] - :param folder: The folder that this data flow is in. If not specified, Data flow will appear at - the root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowFolder - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DataFlowFolder'}, - } - - _subtype_map = { - 'type': {'MappingDataFlow': 'MappingDataFlow'} - } - - def __init__( - self, - *, - description: Optional[str] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DataFlowFolder"] = None, - **kwargs - ): - super(DataFlow, self).__init__(**kwargs) - self.type = None # type: Optional[str] - self.description = description - self.annotations = annotations - self.folder = folder - - -class DataFlowDebugCommandRequest(msrest.serialization.Model): - """Request body structure for data flow expression preview. - - All required parameters must be populated in order to send to Azure. - - :param session_id: Required. The ID of data flow debug session. - :type session_id: str - :param data_flow_name: The data flow which contains the debug session. - :type data_flow_name: str - :param command_name: The command name. - :type command_name: str - :param command_payload: Required. The command payload object. - :type command_payload: any - """ - - _validation = { - 'session_id': {'required': True}, - 'command_payload': {'required': True}, - } - - _attribute_map = { - 'session_id': {'key': 'sessionId', 'type': 'str'}, - 'data_flow_name': {'key': 'dataFlowName', 'type': 'str'}, - 'command_name': {'key': 'commandName', 'type': 'str'}, - 'command_payload': {'key': 'commandPayload', 'type': 'object'}, - } - - def __init__( - self, - *, - session_id: str, - command_payload: Any, - data_flow_name: Optional[str] = None, - command_name: Optional[str] = None, - **kwargs - ): - super(DataFlowDebugCommandRequest, self).__init__(**kwargs) - self.session_id = session_id - self.data_flow_name = data_flow_name - self.command_name = command_name - self.command_payload = command_payload - - -class DataFlowDebugCommandResponse(msrest.serialization.Model): - """Response body structure of data flow result for data preview, statistics or expression preview. - - :param status: The run status of data preview, statistics or expression preview. - :type status: str - :param data: The result data of data preview, statistics or expression preview. - :type data: str - """ - - _attribute_map = { - 'status': {'key': 'status', 'type': 'str'}, - 'data': {'key': 'data', 'type': 'str'}, - } - - def __init__( - self, - *, - status: Optional[str] = None, - data: Optional[str] = None, - **kwargs - ): - super(DataFlowDebugCommandResponse, self).__init__(**kwargs) - self.status = status - self.data = data - - -class DataFlowDebugPackage(msrest.serialization.Model): - """Request body structure for starting data flow debug session. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param session_id: The ID of data flow debug session. - :type session_id: str - :param data_flow: Data flow instance. - :type data_flow: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowDebugResource - :param datasets: List of datasets. - :type datasets: list[~azure.synapse.artifacts.v2020_12_01.models.DatasetDebugResource] - :param linked_services: List of linked services. - :type linked_services: - list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceDebugResource] - :param staging: Staging info for debug session. - :type staging: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowStagingInfo - :param debug_settings: Data flow debug settings. - :type debug_settings: - ~azure.synapse.artifacts.v2020_12_01.models.DataFlowDebugPackageDebugSettings - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'session_id': {'key': 'sessionId', 'type': 'str'}, - 'data_flow': {'key': 'dataFlow', 'type': 'DataFlowDebugResource'}, - 'datasets': {'key': 'datasets', 'type': '[DatasetDebugResource]'}, - 'linked_services': {'key': 'linkedServices', 'type': '[LinkedServiceDebugResource]'}, - 'staging': {'key': 'staging', 'type': 'DataFlowStagingInfo'}, - 'debug_settings': {'key': 'debugSettings', 'type': 'DataFlowDebugPackageDebugSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - session_id: Optional[str] = None, - data_flow: Optional["DataFlowDebugResource"] = None, - datasets: Optional[List["DatasetDebugResource"]] = None, - linked_services: Optional[List["LinkedServiceDebugResource"]] = None, - staging: Optional["DataFlowStagingInfo"] = None, - debug_settings: Optional["DataFlowDebugPackageDebugSettings"] = None, - **kwargs - ): - super(DataFlowDebugPackage, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.session_id = session_id - self.data_flow = data_flow - self.datasets = datasets - self.linked_services = linked_services - self.staging = staging - self.debug_settings = debug_settings - - -class DataFlowDebugPackageDebugSettings(msrest.serialization.Model): - """Data flow debug settings. - - :param source_settings: Source setting for data flow debug. - :type source_settings: list[~azure.synapse.artifacts.v2020_12_01.models.DataFlowSourceSetting] - :param parameters: Data flow parameters. - :type parameters: dict[str, any] - :param dataset_parameters: Parameters for dataset. - :type dataset_parameters: any - """ - - _attribute_map = { - 'source_settings': {'key': 'sourceSettings', 'type': '[DataFlowSourceSetting]'}, - 'parameters': {'key': 'parameters', 'type': '{object}'}, - 'dataset_parameters': {'key': 'datasetParameters', 'type': 'object'}, - } - - def __init__( - self, - *, - source_settings: Optional[List["DataFlowSourceSetting"]] = None, - parameters: Optional[Dict[str, Any]] = None, - dataset_parameters: Optional[Any] = None, - **kwargs - ): - super(DataFlowDebugPackageDebugSettings, self).__init__(**kwargs) - self.source_settings = source_settings - self.parameters = parameters - self.dataset_parameters = dataset_parameters - - -class DataFlowDebugPreviewDataRequest(msrest.serialization.Model): - """Request body structure for data flow preview data. - - :param session_id: The ID of data flow debug session. - :type session_id: str - :param data_flow_name: The data flow which contains the debug session. - :type data_flow_name: str - :param stream_name: The output stream name. - :type stream_name: str - :param row_limits: The row limit for preview request. - :type row_limits: int - """ - - _attribute_map = { - 'session_id': {'key': 'sessionId', 'type': 'str'}, - 'data_flow_name': {'key': 'dataFlowName', 'type': 'str'}, - 'stream_name': {'key': 'streamName', 'type': 'str'}, - 'row_limits': {'key': 'rowLimits', 'type': 'int'}, - } - - def __init__( - self, - *, - session_id: Optional[str] = None, - data_flow_name: Optional[str] = None, - stream_name: Optional[str] = None, - row_limits: Optional[int] = None, - **kwargs - ): - super(DataFlowDebugPreviewDataRequest, self).__init__(**kwargs) - self.session_id = session_id - self.data_flow_name = data_flow_name - self.stream_name = stream_name - self.row_limits = row_limits - - -class DataFlowDebugQueryResponse(msrest.serialization.Model): - """Response body structure of data flow query for data preview, statistics or expression preview. - - :param run_id: The run ID of data flow debug session. - :type run_id: str - """ - - _attribute_map = { - 'run_id': {'key': 'runId', 'type': 'str'}, - } - - def __init__( - self, - *, - run_id: Optional[str] = None, - **kwargs - ): - super(DataFlowDebugQueryResponse, self).__init__(**kwargs) - self.run_id = run_id - - -class SubResourceDebugResource(msrest.serialization.Model): - """Azure Synapse nested debug resource. - - :param name: The resource name. - :type name: str - """ - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - } - - def __init__( - self, - *, - name: Optional[str] = None, - **kwargs - ): - super(SubResourceDebugResource, self).__init__(**kwargs) - self.name = name - - -class DataFlowDebugResource(SubResourceDebugResource): - """Data flow debug resource. - - All required parameters must be populated in order to send to Azure. - - :param name: The resource name. - :type name: str - :param properties: Required. Data flow properties. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.DataFlow - """ - - _validation = { - 'properties': {'required': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'DataFlow'}, - } - - def __init__( - self, - *, - properties: "DataFlow", - name: Optional[str] = None, - **kwargs - ): - super(DataFlowDebugResource, self).__init__(name=name, **kwargs) - self.properties = properties - - -class DataFlowDebugResultResponse(msrest.serialization.Model): - """Response body structure of data flow result for data preview, statistics or expression preview. - - :param status: The run status of data preview, statistics or expression preview. - :type status: str - :param data: The result data of data preview, statistics or expression preview. - :type data: str - """ - - _attribute_map = { - 'status': {'key': 'status', 'type': 'str'}, - 'data': {'key': 'data', 'type': 'str'}, - } - - def __init__( - self, - *, - status: Optional[str] = None, - data: Optional[str] = None, - **kwargs - ): - super(DataFlowDebugResultResponse, self).__init__(**kwargs) - self.status = status - self.data = data - - -class DataFlowDebugSessionInfo(msrest.serialization.Model): - """Data flow debug session info. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param data_flow_name: The name of the data flow. - :type data_flow_name: str - :param compute_type: Compute type of the cluster. - :type compute_type: str - :param core_count: Core count of the cluster. - :type core_count: int - :param node_count: Node count of the cluster. (deprecated property). - :type node_count: int - :param integration_runtime_name: Attached integration runtime name of data flow debug session. - :type integration_runtime_name: str - :param session_id: The ID of data flow debug session. - :type session_id: str - :param start_time: Start time of data flow debug session. - :type start_time: str - :param time_to_live_in_minutes: Compute type of the cluster. - :type time_to_live_in_minutes: int - :param last_activity_time: Last activity time of data flow debug session. - :type last_activity_time: str - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'data_flow_name': {'key': 'dataFlowName', 'type': 'str'}, - 'compute_type': {'key': 'computeType', 'type': 'str'}, - 'core_count': {'key': 'coreCount', 'type': 'int'}, - 'node_count': {'key': 'nodeCount', 'type': 'int'}, - 'integration_runtime_name': {'key': 'integrationRuntimeName', 'type': 'str'}, - 'session_id': {'key': 'sessionId', 'type': 'str'}, - 'start_time': {'key': 'startTime', 'type': 'str'}, - 'time_to_live_in_minutes': {'key': 'timeToLiveInMinutes', 'type': 'int'}, - 'last_activity_time': {'key': 'lastActivityTime', 'type': 'str'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - data_flow_name: Optional[str] = None, - compute_type: Optional[str] = None, - core_count: Optional[int] = None, - node_count: Optional[int] = None, - integration_runtime_name: Optional[str] = None, - session_id: Optional[str] = None, - start_time: Optional[str] = None, - time_to_live_in_minutes: Optional[int] = None, - last_activity_time: Optional[str] = None, - **kwargs - ): - super(DataFlowDebugSessionInfo, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.data_flow_name = data_flow_name - self.compute_type = compute_type - self.core_count = core_count - self.node_count = node_count - self.integration_runtime_name = integration_runtime_name - self.session_id = session_id - self.start_time = start_time - self.time_to_live_in_minutes = time_to_live_in_minutes - self.last_activity_time = last_activity_time - - -class DataFlowDebugStatisticsRequest(msrest.serialization.Model): - """Request body structure for data flow statistics. - - :param session_id: The ID of data flow debug session. - :type session_id: str - :param data_flow_name: The data flow which contains the debug session. - :type data_flow_name: str - :param stream_name: The output stream name. - :type stream_name: str - :param columns: List of column names. - :type columns: list[str] - """ - - _attribute_map = { - 'session_id': {'key': 'sessionId', 'type': 'str'}, - 'data_flow_name': {'key': 'dataFlowName', 'type': 'str'}, - 'stream_name': {'key': 'streamName', 'type': 'str'}, - 'columns': {'key': 'columns', 'type': '[str]'}, - } - - def __init__( - self, - *, - session_id: Optional[str] = None, - data_flow_name: Optional[str] = None, - stream_name: Optional[str] = None, - columns: Optional[List[str]] = None, - **kwargs - ): - super(DataFlowDebugStatisticsRequest, self).__init__(**kwargs) - self.session_id = session_id - self.data_flow_name = data_flow_name - self.stream_name = stream_name - self.columns = columns - - -class DataFlowFolder(msrest.serialization.Model): - """The folder that this data flow is in. If not specified, Data flow will appear at the root level. - - :param name: The name of the folder that this data flow is in. - :type name: str - """ - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - } - - def __init__( - self, - *, - name: Optional[str] = None, - **kwargs - ): - super(DataFlowFolder, self).__init__(**kwargs) - self.name = name - - -class DataFlowListResponse(msrest.serialization.Model): - """A list of data flow resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of data flows. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.DataFlowResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[DataFlowResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - *, - value: List["DataFlowResource"], - next_link: Optional[str] = None, - **kwargs - ): - super(DataFlowListResponse, self).__init__(**kwargs) - self.value = value - self.next_link = next_link - - -class DataFlowReference(msrest.serialization.Model): - """Data flow reference type. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Data flow reference type. Possible values include: "DataFlowReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.DataFlowReferenceType - :param reference_name: Required. Reference data flow name. - :type reference_name: str - :param dataset_parameters: Reference data flow parameters from dataset. - :type dataset_parameters: any - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - 'dataset_parameters': {'key': 'datasetParameters', 'type': 'object'}, - } - - def __init__( - self, - *, - type: Union[str, "DataFlowReferenceType"], - reference_name: str, - additional_properties: Optional[Dict[str, Any]] = None, - dataset_parameters: Optional[Any] = None, - **kwargs - ): - super(DataFlowReference, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = type - self.reference_name = reference_name - self.dataset_parameters = dataset_parameters - - -class SubResource(AzureEntityResource): - """Azure Synapse nested resource, which belongs to a workspace. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(SubResource, self).__init__(**kwargs) - - -class DataFlowResource(SubResource): - """Data flow resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Data flow properties. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.DataFlow - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'DataFlow'}, - } - - def __init__( - self, - *, - properties: "DataFlow", - **kwargs - ): - super(DataFlowResource, self).__init__(**kwargs) - self.properties = properties - - -class Transformation(msrest.serialization.Model): - """A data flow transformation. - - All required parameters must be populated in order to send to Azure. - - :param name: Required. Transformation name. - :type name: str - :param description: Transformation description. - :type description: str - """ - - _validation = { - 'name': {'required': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - } - - def __init__( - self, - *, - name: str, - description: Optional[str] = None, - **kwargs - ): - super(Transformation, self).__init__(**kwargs) - self.name = name - self.description = description - - -class DataFlowSink(Transformation): - """Transformation for data flow sink. - - All required parameters must be populated in order to send to Azure. - - :param name: Required. Transformation name. - :type name: str - :param description: Transformation description. - :type description: str - :param dataset: Dataset reference. - :type dataset: ~azure.synapse.artifacts.v2020_12_01.models.DatasetReference - :param linked_service: Linked service reference. - :type linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param schema_linked_service: Schema linked service reference. - :type schema_linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - """ - - _validation = { - 'name': {'required': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'dataset': {'key': 'dataset', 'type': 'DatasetReference'}, - 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, - 'schema_linked_service': {'key': 'schemaLinkedService', 'type': 'LinkedServiceReference'}, - } - - def __init__( - self, - *, - name: str, - description: Optional[str] = None, - dataset: Optional["DatasetReference"] = None, - linked_service: Optional["LinkedServiceReference"] = None, - schema_linked_service: Optional["LinkedServiceReference"] = None, - **kwargs - ): - super(DataFlowSink, self).__init__(name=name, description=description, **kwargs) - self.dataset = dataset - self.linked_service = linked_service - self.schema_linked_service = schema_linked_service - - -class DataFlowSource(Transformation): - """Transformation for data flow source. - - All required parameters must be populated in order to send to Azure. - - :param name: Required. Transformation name. - :type name: str - :param description: Transformation description. - :type description: str - :param dataset: Dataset reference. - :type dataset: ~azure.synapse.artifacts.v2020_12_01.models.DatasetReference - :param linked_service: Linked service reference. - :type linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param schema_linked_service: Schema linked service reference. - :type schema_linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - """ - - _validation = { - 'name': {'required': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'dataset': {'key': 'dataset', 'type': 'DatasetReference'}, - 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, - 'schema_linked_service': {'key': 'schemaLinkedService', 'type': 'LinkedServiceReference'}, - } - - def __init__( - self, - *, - name: str, - description: Optional[str] = None, - dataset: Optional["DatasetReference"] = None, - linked_service: Optional["LinkedServiceReference"] = None, - schema_linked_service: Optional["LinkedServiceReference"] = None, - **kwargs - ): - super(DataFlowSource, self).__init__(name=name, description=description, **kwargs) - self.dataset = dataset - self.linked_service = linked_service - self.schema_linked_service = schema_linked_service - - -class DataFlowSourceSetting(msrest.serialization.Model): - """Definition of data flow source setting for debug. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param source_name: The data flow source name. - :type source_name: str - :param row_limit: Defines the row limit of data flow source in debug. - :type row_limit: int - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'source_name': {'key': 'sourceName', 'type': 'str'}, - 'row_limit': {'key': 'rowLimit', 'type': 'int'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_name: Optional[str] = None, - row_limit: Optional[int] = None, - **kwargs - ): - super(DataFlowSourceSetting, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.source_name = source_name - self.row_limit = row_limit - - -class DataFlowStagingInfo(msrest.serialization.Model): - """Staging info for execute data flow activity. - - :param linked_service: Staging linked service reference. - :type linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param folder_path: Folder path for staging blob. - :type folder_path: str - """ - - _attribute_map = { - 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, - 'folder_path': {'key': 'folderPath', 'type': 'str'}, - } - - def __init__( - self, - *, - linked_service: Optional["LinkedServiceReference"] = None, - folder_path: Optional[str] = None, - **kwargs - ): - super(DataFlowStagingInfo, self).__init__(**kwargs) - self.linked_service = linked_service - self.folder_path = folder_path - - -class DataLakeAnalyticsUSQLActivity(ExecutionActivity): - """Data Lake Analytics U-SQL activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param script_path: Required. Case-sensitive path to folder that contains the U-SQL script. - Type: string (or Expression with resultType string). - :type script_path: any - :param script_linked_service: Required. Script linked service reference. - :type script_linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param degree_of_parallelism: The maximum number of nodes simultaneously used to run the job. - Default value is 1. Type: integer (or Expression with resultType integer), minimum: 1. - :type degree_of_parallelism: any - :param priority: Determines which jobs out of all that are queued should be selected to run - first. The lower the number, the higher the priority. Default value is 1000. Type: integer (or - Expression with resultType integer), minimum: 1. - :type priority: any - :param parameters: Parameters for U-SQL job request. - :type parameters: dict[str, any] - :param runtime_version: Runtime version of the U-SQL engine to use. Type: string (or Expression - with resultType string). - :type runtime_version: any - :param compilation_mode: Compilation mode of U-SQL. Must be one of these values : Semantic, - Full and SingleBox. Type: string (or Expression with resultType string). - :type compilation_mode: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'script_path': {'required': True}, - 'script_linked_service': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'script_path': {'key': 'typeProperties.scriptPath', 'type': 'object'}, - 'script_linked_service': {'key': 'typeProperties.scriptLinkedService', 'type': 'LinkedServiceReference'}, - 'degree_of_parallelism': {'key': 'typeProperties.degreeOfParallelism', 'type': 'object'}, - 'priority': {'key': 'typeProperties.priority', 'type': 'object'}, - 'parameters': {'key': 'typeProperties.parameters', 'type': '{object}'}, - 'runtime_version': {'key': 'typeProperties.runtimeVersion', 'type': 'object'}, - 'compilation_mode': {'key': 'typeProperties.compilationMode', 'type': 'object'}, - } - - def __init__( - self, - *, - name: str, - script_path: Any, - script_linked_service: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - degree_of_parallelism: Optional[Any] = None, - priority: Optional[Any] = None, - parameters: Optional[Dict[str, Any]] = None, - runtime_version: Optional[Any] = None, - compilation_mode: Optional[Any] = None, - **kwargs - ): - super(DataLakeAnalyticsUSQLActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'DataLakeAnalyticsU-SQL' # type: str - self.script_path = script_path - self.script_linked_service = script_linked_service - self.degree_of_parallelism = degree_of_parallelism - self.priority = priority - self.parameters = parameters - self.runtime_version = runtime_version - self.compilation_mode = compilation_mode - - -class DataLakeStorageAccountDetails(msrest.serialization.Model): - """Details of the data lake storage account associated with the workspace. - - :param account_url: Account URL. - :type account_url: str - :param filesystem: Filesystem name. - :type filesystem: str - """ - - _attribute_map = { - 'account_url': {'key': 'accountUrl', 'type': 'str'}, - 'filesystem': {'key': 'filesystem', 'type': 'str'}, - } - - def __init__( - self, - *, - account_url: Optional[str] = None, - filesystem: Optional[str] = None, - **kwargs - ): - super(DataLakeStorageAccountDetails, self).__init__(**kwargs) - self.account_url = account_url - self.filesystem = filesystem - - -class DatasetCompression(msrest.serialization.Model): - """The compression method used on a dataset. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: DatasetBZip2Compression, DatasetDeflateCompression, DatasetGZipCompression, DatasetTarCompression, DatasetTarGZipCompression, DatasetZipDeflateCompression. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, - } - - _subtype_map = { - 'type': {'BZip2': 'DatasetBZip2Compression', 'Deflate': 'DatasetDeflateCompression', 'GZip': 'DatasetGZipCompression', 'Tar': 'DatasetTarCompression', 'TarGZip': 'DatasetTarGZipCompression', 'ZipDeflate': 'DatasetZipDeflateCompression'} - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(DatasetCompression, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = 'DatasetCompression' # type: str - - -class DatasetBZip2Compression(DatasetCompression): - """The BZip2 compression method used on a dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(DatasetBZip2Compression, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'BZip2' # type: str - - -class DatasetDataElement(msrest.serialization.Model): - """Columns that define the structure of the dataset. - - :param name: Name of the column. Type: string (or Expression with resultType string). - :type name: any - :param type: Type of the column. Type: string (or Expression with resultType string). - :type type: any - """ - - _attribute_map = { - 'name': {'key': 'name', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'object'}, - } - - def __init__( - self, - *, - name: Optional[Any] = None, - type: Optional[Any] = None, - **kwargs - ): - super(DatasetDataElement, self).__init__(**kwargs) - self.name = name - self.type = type - - -class DatasetDebugResource(SubResourceDebugResource): - """Dataset debug resource. - - All required parameters must be populated in order to send to Azure. - - :param name: The resource name. - :type name: str - :param properties: Required. Dataset properties. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.Dataset - """ - - _validation = { - 'properties': {'required': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'Dataset'}, - } - - def __init__( - self, - *, - properties: "Dataset", - name: Optional[str] = None, - **kwargs - ): - super(DatasetDebugResource, self).__init__(name=name, **kwargs) - self.properties = properties - - -class DatasetDeflateCompression(DatasetCompression): - """The Deflate compression method used on a dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - :param level: The Deflate compression level. - :type level: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, - 'level': {'key': 'level', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - level: Optional[Any] = None, - **kwargs - ): - super(DatasetDeflateCompression, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'Deflate' # type: str - self.level = level - - -class DatasetFolder(msrest.serialization.Model): - """The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - - :param name: The name of the folder that this Dataset is in. - :type name: str - """ - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - } - - def __init__( - self, - *, - name: Optional[str] = None, - **kwargs - ): - super(DatasetFolder, self).__init__(**kwargs) - self.name = name - - -class DatasetGZipCompression(DatasetCompression): - """The GZip compression method used on a dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - :param level: The GZip compression level. - :type level: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, - 'level': {'key': 'level', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - level: Optional[Any] = None, - **kwargs - ): - super(DatasetGZipCompression, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'GZip' # type: str - self.level = level - - -class DatasetListResponse(msrest.serialization.Model): - """A list of dataset resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of datasets. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.DatasetResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[DatasetResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - *, - value: List["DatasetResource"], - next_link: Optional[str] = None, - **kwargs - ): - super(DatasetListResponse, self).__init__(**kwargs) - self.value = value - self.next_link = next_link - - -class DatasetReference(msrest.serialization.Model): - """Dataset reference type. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Dataset reference type. Possible values include: "DatasetReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.DatasetReferenceType - :param reference_name: Required. Reference dataset name. - :type reference_name: str - :param parameters: Arguments for dataset. - :type parameters: dict[str, any] - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{object}'}, - } - - def __init__( - self, - *, - type: Union[str, "DatasetReferenceType"], - reference_name: str, - parameters: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(DatasetReference, self).__init__(**kwargs) - self.type = type - self.reference_name = reference_name - self.parameters = parameters - - -class DatasetResource(SubResource): - """Dataset resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Dataset properties. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.Dataset - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'Dataset'}, - } - - def __init__( - self, - *, - properties: "Dataset", - **kwargs - ): - super(DatasetResource, self).__init__(**kwargs) - self.properties = properties - - -class DatasetSchemaDataElement(msrest.serialization.Model): - """Columns that define the physical type schema of the dataset. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Name of the schema column. Type: string (or Expression with resultType string). - :type name: any - :param type: Type of the schema column. Type: string (or Expression with resultType string). - :type type: any - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - name: Optional[Any] = None, - type: Optional[Any] = None, - **kwargs - ): - super(DatasetSchemaDataElement, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.name = name - self.type = type - - -class DatasetTarCompression(DatasetCompression): - """The Tar archive method used on a dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(DatasetTarCompression, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'Tar' # type: str - - -class DatasetTarGZipCompression(DatasetCompression): - """The TarGZip compression method used on a dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - :param level: The TarGZip compression level. - :type level: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, - 'level': {'key': 'level', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - level: Optional[Any] = None, - **kwargs - ): - super(DatasetTarGZipCompression, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'TarGZip' # type: str - self.level = level - - -class DatasetZipDeflateCompression(DatasetCompression): - """The ZipDeflate compression method used on a dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression. Type: string (or Expression with resultType - string).Constant filled by server. - :type type: any - :param level: The ZipDeflate compression level. - :type level: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'object'}, - 'level': {'key': 'level', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - level: Optional[Any] = None, - **kwargs - ): - super(DatasetZipDeflateCompression, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'ZipDeflate' # type: str - self.level = level - - -class Db2LinkedService(LinkedService): - """Linked service for DB2 data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: The connection string. It is mutually exclusive with server, - database, authenticationType, userName, packageCollection and certificateCommonName property. - Type: string, SecureString or AzureKeyVaultSecretReference. - :type connection_string: any - :param server: Required. Server name for connection. It is mutually exclusive with - connectionString property. Type: string (or Expression with resultType string). - :type server: any - :param database: Required. Database name for connection. It is mutually exclusive with - connectionString property. Type: string (or Expression with resultType string). - :type database: any - :param authentication_type: AuthenticationType to be used for connection. It is mutually - exclusive with connectionString property. Possible values include: "Basic". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.Db2AuthenticationType - :param username: Username for authentication. It is mutually exclusive with connectionString - property. Type: string (or Expression with resultType string). - :type username: any - :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param package_collection: Under where packages are created when querying database. It is - mutually exclusive with connectionString property. Type: string (or Expression with resultType - string). - :type package_collection: any - :param certificate_common_name: Certificate Common Name when TLS is enabled. It is mutually - exclusive with connectionString property. Type: string (or Expression with resultType string). - :type certificate_common_name: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. It is mutually exclusive with - connectionString property. Type: string (or Expression with resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'server': {'required': True}, - 'database': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'package_collection': {'key': 'typeProperties.packageCollection', 'type': 'object'}, - 'certificate_common_name': {'key': 'typeProperties.certificateCommonName', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - server: Any, - database: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_string: Optional[Any] = None, - authentication_type: Optional[Union[str, "Db2AuthenticationType"]] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - package_collection: Optional[Any] = None, - certificate_common_name: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(Db2LinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Db2' # type: str - self.connection_string = connection_string - self.server = server - self.database = database - self.authentication_type = authentication_type - self.username = username - self.password = password - self.package_collection = package_collection - self.certificate_common_name = certificate_common_name - self.encrypted_credential = encrypted_credential - - -class Db2Source(TabularSource): - """A copy activity source for Db2 databases. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(Db2Source, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'Db2Source' # type: str - self.query = query - - -class Db2TableDataset(Dataset): - """The Db2 table dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param schema_type_properties_schema: The Db2 schema name. Type: string (or Expression with - resultType string). - :type schema_type_properties_schema: any - :param table: The Db2 table name. Type: string (or Expression with resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - schema_type_properties_schema: Optional[Any] = None, - table: Optional[Any] = None, - **kwargs - ): - super(Db2TableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'Db2Table' # type: str - self.table_name = table_name - self.schema_type_properties_schema = schema_type_properties_schema - self.table = table - - -class DeleteActivity(ExecutionActivity): - """Delete activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param recursive: If true, files or sub-folders under current folder path will be deleted - recursively. Default is false. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param max_concurrent_connections: The max concurrent connections to connect data source at the - same time. - :type max_concurrent_connections: int - :param enable_logging: Whether to record detailed logs of delete-activity execution. Default - value is false. Type: boolean (or Expression with resultType boolean). - :type enable_logging: any - :param log_storage_settings: Log storage settings customer need to provide when enableLogging - is true. - :type log_storage_settings: ~azure.synapse.artifacts.v2020_12_01.models.LogStorageSettings - :param dataset: Required. Delete activity dataset reference. - :type dataset: ~azure.synapse.artifacts.v2020_12_01.models.DatasetReference - :param store_settings: Delete activity store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'max_concurrent_connections': {'minimum': 1}, - 'dataset': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'recursive': {'key': 'typeProperties.recursive', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'typeProperties.maxConcurrentConnections', 'type': 'int'}, - 'enable_logging': {'key': 'typeProperties.enableLogging', 'type': 'object'}, - 'log_storage_settings': {'key': 'typeProperties.logStorageSettings', 'type': 'LogStorageSettings'}, - 'dataset': {'key': 'typeProperties.dataset', 'type': 'DatasetReference'}, - 'store_settings': {'key': 'typeProperties.storeSettings', 'type': 'StoreReadSettings'}, - } - - def __init__( - self, - *, - name: str, - dataset: "DatasetReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - recursive: Optional[Any] = None, - max_concurrent_connections: Optional[int] = None, - enable_logging: Optional[Any] = None, - log_storage_settings: Optional["LogStorageSettings"] = None, - store_settings: Optional["StoreReadSettings"] = None, - **kwargs - ): - super(DeleteActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'Delete' # type: str - self.recursive = recursive - self.max_concurrent_connections = max_concurrent_connections - self.enable_logging = enable_logging - self.log_storage_settings = log_storage_settings - self.dataset = dataset - self.store_settings = store_settings - - -class DeleteDataFlowDebugSessionRequest(msrest.serialization.Model): - """Request body structure for deleting data flow debug session. - - :param session_id: The ID of data flow debug session. - :type session_id: str - :param data_flow_name: The data flow which contains the debug session. - :type data_flow_name: str - """ - - _attribute_map = { - 'session_id': {'key': 'sessionId', 'type': 'str'}, - 'data_flow_name': {'key': 'dataFlowName', 'type': 'str'}, - } - - def __init__( - self, - *, - session_id: Optional[str] = None, - data_flow_name: Optional[str] = None, - **kwargs - ): - super(DeleteDataFlowDebugSessionRequest, self).__init__(**kwargs) - self.session_id = session_id - self.data_flow_name = data_flow_name - - -class DelimitedTextDataset(Dataset): - """Delimited text dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param location: The location of the delimited text storage. - :type location: ~azure.synapse.artifacts.v2020_12_01.models.DatasetLocation - :param column_delimiter: The column delimiter. Type: string (or Expression with resultType - string). - :type column_delimiter: any - :param row_delimiter: The row delimiter. Type: string (or Expression with resultType string). - :type row_delimiter: any - :param encoding_name: The code page name of the preferred encoding. If miss, the default value - is UTF-8, unless BOM denotes another Unicode encoding. Refer to the name column of the table in - the following link to set supported values: - https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string (or Expression with - resultType string). - :type encoding_name: any - :param compression_codec: Possible values include: "bzip2", "gzip", "deflate", "zipDeflate", - "snappy", "lz4", "tar", "tarGZip". - :type compression_codec: str or ~azure.synapse.artifacts.v2020_12_01.models.CompressionCodec - :param compression_level: The data compression method used for DelimitedText. - :type compression_level: any - :param quote_char: The quote character. Type: string (or Expression with resultType string). - :type quote_char: any - :param escape_char: The escape character. Type: string (or Expression with resultType string). - :type escape_char: any - :param first_row_as_header: When used as input, treat the first row of data as headers. When - used as output,write the headers into the output as the first row of data. The default value is - false. Type: boolean (or Expression with resultType boolean). - :type first_row_as_header: any - :param null_value: The null value string. Type: string (or Expression with resultType string). - :type null_value: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'location': {'key': 'typeProperties.location', 'type': 'DatasetLocation'}, - 'column_delimiter': {'key': 'typeProperties.columnDelimiter', 'type': 'object'}, - 'row_delimiter': {'key': 'typeProperties.rowDelimiter', 'type': 'object'}, - 'encoding_name': {'key': 'typeProperties.encodingName', 'type': 'object'}, - 'compression_codec': {'key': 'typeProperties.compressionCodec', 'type': 'str'}, - 'compression_level': {'key': 'typeProperties.compressionLevel', 'type': 'object'}, - 'quote_char': {'key': 'typeProperties.quoteChar', 'type': 'object'}, - 'escape_char': {'key': 'typeProperties.escapeChar', 'type': 'object'}, - 'first_row_as_header': {'key': 'typeProperties.firstRowAsHeader', 'type': 'object'}, - 'null_value': {'key': 'typeProperties.nullValue', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - location: Optional["DatasetLocation"] = None, - column_delimiter: Optional[Any] = None, - row_delimiter: Optional[Any] = None, - encoding_name: Optional[Any] = None, - compression_codec: Optional[Union[str, "CompressionCodec"]] = None, - compression_level: Optional[Any] = None, - quote_char: Optional[Any] = None, - escape_char: Optional[Any] = None, - first_row_as_header: Optional[Any] = None, - null_value: Optional[Any] = None, - **kwargs - ): - super(DelimitedTextDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'DelimitedText' # type: str - self.location = location - self.column_delimiter = column_delimiter - self.row_delimiter = row_delimiter - self.encoding_name = encoding_name - self.compression_codec = compression_codec - self.compression_level = compression_level - self.quote_char = quote_char - self.escape_char = escape_char - self.first_row_as_header = first_row_as_header - self.null_value = null_value - - -class DelimitedTextReadSettings(FormatReadSettings): - """Delimited text read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param skip_line_count: Indicates the number of non-empty rows to skip when reading data from - input files. Type: integer (or Expression with resultType integer). - :type skip_line_count: any - :param compression_properties: Compression settings. - :type compression_properties: - ~azure.synapse.artifacts.v2020_12_01.models.CompressionReadSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'skip_line_count': {'key': 'skipLineCount', 'type': 'object'}, - 'compression_properties': {'key': 'compressionProperties', 'type': 'CompressionReadSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - skip_line_count: Optional[Any] = None, - compression_properties: Optional["CompressionReadSettings"] = None, - **kwargs - ): - super(DelimitedTextReadSettings, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'DelimitedTextReadSettings' # type: str - self.skip_line_count = skip_line_count - self.compression_properties = compression_properties - - -class DelimitedTextSink(CopySink): - """A copy activity DelimitedText sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: DelimitedText store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreWriteSettings - :param format_settings: DelimitedText format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.DelimitedTextWriteSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreWriteSettings'}, - 'format_settings': {'key': 'formatSettings', 'type': 'DelimitedTextWriteSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - store_settings: Optional["StoreWriteSettings"] = None, - format_settings: Optional["DelimitedTextWriteSettings"] = None, - **kwargs - ): - super(DelimitedTextSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'DelimitedTextSink' # type: str - self.store_settings = store_settings - self.format_settings = format_settings - - -class DelimitedTextSource(CopySource): - """A copy activity DelimitedText source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: DelimitedText store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - :param format_settings: DelimitedText format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.DelimitedTextReadSettings - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'format_settings': {'key': 'formatSettings', 'type': 'DelimitedTextReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - store_settings: Optional["StoreReadSettings"] = None, - format_settings: Optional["DelimitedTextReadSettings"] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(DelimitedTextSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'DelimitedTextSource' # type: str - self.store_settings = store_settings - self.format_settings = format_settings - self.additional_columns = additional_columns - - -class DelimitedTextWriteSettings(FormatWriteSettings): - """Delimited text write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param quote_all_text: Indicates whether string values should always be enclosed with quotes. - Type: boolean (or Expression with resultType boolean). - :type quote_all_text: any - :param file_extension: Required. The file extension used to create the files. Type: string (or - Expression with resultType string). - :type file_extension: any - :param max_rows_per_file: Limit the written file's row count to be smaller than or equal to the - specified count. Type: integer (or Expression with resultType integer). - :type max_rows_per_file: any - :param file_name_prefix: Specifies the file name pattern - :code:``_:code:``.:code:`` when copy from non-file - based store without partitionOptions. Type: string (or Expression with resultType string). - :type file_name_prefix: any - """ - - _validation = { - 'type': {'required': True}, - 'file_extension': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'quote_all_text': {'key': 'quoteAllText', 'type': 'object'}, - 'file_extension': {'key': 'fileExtension', 'type': 'object'}, - 'max_rows_per_file': {'key': 'maxRowsPerFile', 'type': 'object'}, - 'file_name_prefix': {'key': 'fileNamePrefix', 'type': 'object'}, - } - - def __init__( - self, - *, - file_extension: Any, - additional_properties: Optional[Dict[str, Any]] = None, - quote_all_text: Optional[Any] = None, - max_rows_per_file: Optional[Any] = None, - file_name_prefix: Optional[Any] = None, - **kwargs - ): - super(DelimitedTextWriteSettings, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'DelimitedTextWriteSettings' # type: str - self.quote_all_text = quote_all_text - self.file_extension = file_extension - self.max_rows_per_file = max_rows_per_file - self.file_name_prefix = file_name_prefix - - -class DependencyReference(msrest.serialization.Model): - """Referenced dependency. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: SelfDependencyTumblingWindowTriggerReference, TriggerDependencyReference. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. The type of dependency reference.Constant filled by server. - :type type: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - } - - _subtype_map = { - 'type': {'SelfDependencyTumblingWindowTriggerReference': 'SelfDependencyTumblingWindowTriggerReference', 'TriggerDependencyReference': 'TriggerDependencyReference'} - } - - def __init__( - self, - **kwargs - ): - super(DependencyReference, self).__init__(**kwargs) - self.type = None # type: Optional[str] - - -class DistcpSettings(msrest.serialization.Model): - """Distcp settings. - - All required parameters must be populated in order to send to Azure. - - :param resource_manager_endpoint: Required. Specifies the Yarn ResourceManager endpoint. Type: - string (or Expression with resultType string). - :type resource_manager_endpoint: any - :param temp_script_path: Required. Specifies an existing folder path which will be used to - store temp Distcp command script. The script file is generated by ADF and will be removed after - Copy job finished. Type: string (or Expression with resultType string). - :type temp_script_path: any - :param distcp_options: Specifies the Distcp options. Type: string (or Expression with - resultType string). - :type distcp_options: any - """ - - _validation = { - 'resource_manager_endpoint': {'required': True}, - 'temp_script_path': {'required': True}, - } - - _attribute_map = { - 'resource_manager_endpoint': {'key': 'resourceManagerEndpoint', 'type': 'object'}, - 'temp_script_path': {'key': 'tempScriptPath', 'type': 'object'}, - 'distcp_options': {'key': 'distcpOptions', 'type': 'object'}, - } - - def __init__( - self, - *, - resource_manager_endpoint: Any, - temp_script_path: Any, - distcp_options: Optional[Any] = None, - **kwargs - ): - super(DistcpSettings, self).__init__(**kwargs) - self.resource_manager_endpoint = resource_manager_endpoint - self.temp_script_path = temp_script_path - self.distcp_options = distcp_options - - -class DocumentDbCollectionDataset(Dataset): - """Microsoft Azure Document Database Collection dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param collection_name: Required. Document Database collection name. Type: string (or - Expression with resultType string). - :type collection_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'collection_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'collection_name': {'key': 'typeProperties.collectionName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - collection_name: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - **kwargs - ): - super(DocumentDbCollectionDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'DocumentDbCollection' # type: str - self.collection_name = collection_name - - -class DocumentDbCollectionSink(CopySink): - """A copy activity Document Database Collection sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param nesting_separator: Nested properties separator. Default is . (dot). Type: string (or - Expression with resultType string). - :type nesting_separator: any - :param write_behavior: Describes how to write data to Azure Cosmos DB. Type: string (or - Expression with resultType string). Allowed values: insert and upsert. - :type write_behavior: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'nesting_separator': {'key': 'nestingSeparator', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - nesting_separator: Optional[Any] = None, - write_behavior: Optional[Any] = None, - **kwargs - ): - super(DocumentDbCollectionSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'DocumentDbCollectionSink' # type: str - self.nesting_separator = nesting_separator - self.write_behavior = write_behavior - - -class DocumentDbCollectionSource(CopySource): - """A copy activity Document Database Collection source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: Documents query. Type: string (or Expression with resultType string). - :type query: any - :param nesting_separator: Nested properties separator. Type: string (or Expression with - resultType string). - :type nesting_separator: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'nesting_separator': {'key': 'nestingSeparator', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query: Optional[Any] = None, - nesting_separator: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(DocumentDbCollectionSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'DocumentDbCollectionSource' # type: str - self.query = query - self.nesting_separator = nesting_separator - self.query_timeout = query_timeout - self.additional_columns = additional_columns - - -class DrillLinkedService(LinkedService): - """Drill server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: An ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_string: Optional[Any] = None, - pwd: Optional["AzureKeyVaultSecretReference"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(DrillLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Drill' # type: str - self.connection_string = connection_string - self.pwd = pwd - self.encrypted_credential = encrypted_credential - - -class DrillSource(TabularSource): - """A copy activity Drill server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(DrillSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'DrillSource' # type: str - self.query = query - - -class DrillTableDataset(Dataset): - """Drill server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The table name of the Drill. Type: string (or Expression with resultType string). - :type table: any - :param schema_type_properties_schema: The schema name of the Drill. Type: string (or Expression - with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - table: Optional[Any] = None, - schema_type_properties_schema: Optional[Any] = None, - **kwargs - ): - super(DrillTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'DrillTable' # type: str - self.table_name = table_name - self.table = table - self.schema_type_properties_schema = schema_type_properties_schema - - -class DWCopyCommandDefaultValue(msrest.serialization.Model): - """Default value. - - :param column_name: Column name. Type: object (or Expression with resultType string). - :type column_name: any - :param default_value: The default value of the column. Type: object (or Expression with - resultType string). - :type default_value: any - """ - - _attribute_map = { - 'column_name': {'key': 'columnName', 'type': 'object'}, - 'default_value': {'key': 'defaultValue', 'type': 'object'}, - } - - def __init__( - self, - *, - column_name: Optional[Any] = None, - default_value: Optional[Any] = None, - **kwargs - ): - super(DWCopyCommandDefaultValue, self).__init__(**kwargs) - self.column_name = column_name - self.default_value = default_value - - -class DWCopyCommandSettings(msrest.serialization.Model): - """DW Copy Command settings. - - :param default_values: Specifies the default values for each target column in SQL DW. The - default values in the property overwrite the DEFAULT constraint set in the DB, and identity - column cannot have a default value. Type: array of objects (or Expression with resultType array - of objects). - :type default_values: - list[~azure.synapse.artifacts.v2020_12_01.models.DWCopyCommandDefaultValue] - :param additional_options: Additional options directly passed to SQL DW in Copy Command. Type: - key value pairs (value should be string type) (or Expression with resultType object). Example: - "additionalOptions": { "MAXERRORS": "1000", "DATEFORMAT": "'ymd'" }. - :type additional_options: dict[str, str] - """ - - _attribute_map = { - 'default_values': {'key': 'defaultValues', 'type': '[DWCopyCommandDefaultValue]'}, - 'additional_options': {'key': 'additionalOptions', 'type': '{str}'}, - } - - def __init__( - self, - *, - default_values: Optional[List["DWCopyCommandDefaultValue"]] = None, - additional_options: Optional[Dict[str, str]] = None, - **kwargs - ): - super(DWCopyCommandSettings, self).__init__(**kwargs) - self.default_values = default_values - self.additional_options = additional_options - - -class DynamicExecutorAllocation(msrest.serialization.Model): - """Dynamic Executor Allocation Properties. - - :param enabled: Indicates whether Dynamic Executor Allocation is enabled or not. - :type enabled: bool - """ - - _attribute_map = { - 'enabled': {'key': 'enabled', 'type': 'bool'}, - } - - def __init__( - self, - *, - enabled: Optional[bool] = None, - **kwargs - ): - super(DynamicExecutorAllocation, self).__init__(**kwargs) - self.enabled = enabled - - -class DynamicsAXLinkedService(LinkedService): - """Dynamics AX linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param url: Required. The Dynamics AX (or Dynamics 365 Finance and Operations) instance OData - endpoint. - :type url: any - :param service_principal_id: Required. Specify the application's client ID. Type: string (or - Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: Required. Specify the application's key. Mark this field as a - SecureString to store it securely in Data Factory, or reference a secret stored in Azure Key - Vault. Type: string (or Expression with resultType string). - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: Required. Specify the tenant information (domain name or tenant ID) under which - your application resides. Retrieve it by hovering the mouse in the top-right corner of the - Azure portal. Type: string (or Expression with resultType string). - :type tenant: any - :param aad_resource_id: Required. Specify the resource you are requesting authorization. Type: - string (or Expression with resultType string). - :type aad_resource_id: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - 'service_principal_id': {'required': True}, - 'service_principal_key': {'required': True}, - 'tenant': {'required': True}, - 'aad_resource_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'aad_resource_id': {'key': 'typeProperties.aadResourceId', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - url: Any, - service_principal_id: Any, - service_principal_key: "SecretBase", - tenant: Any, - aad_resource_id: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(DynamicsAXLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'DynamicsAX' # type: str - self.url = url - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.aad_resource_id = aad_resource_id - self.encrypted_credential = encrypted_credential - - -class DynamicsAXResourceDataset(Dataset): - """The path of the Dynamics AX OData entity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param path: Required. The path of the Dynamics AX OData entity. Type: string (or Expression - with resultType string). - :type path: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'path': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'path': {'key': 'typeProperties.path', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - path: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - **kwargs - ): - super(DynamicsAXResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'DynamicsAXResource' # type: str - self.path = path - - -class DynamicsAXSource(TabularSource): - """A copy activity Dynamics AX source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - :param http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout - to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - http_request_timeout: Optional[Any] = None, - **kwargs - ): - super(DynamicsAXSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'DynamicsAXSource' # type: str - self.query = query - self.http_request_timeout = http_request_timeout - - -class DynamicsCrmEntityDataset(Dataset): - """The Dynamics CRM entity dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param entity_name: The logical name of the entity. Type: string (or Expression with resultType - string). - :type entity_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'entity_name': {'key': 'typeProperties.entityName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - entity_name: Optional[Any] = None, - **kwargs - ): - super(DynamicsCrmEntityDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'DynamicsCrmEntity' # type: str - self.entity_name = entity_name - - -class DynamicsCrmLinkedService(LinkedService): - """Dynamics CRM linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param deployment_type: Required. The deployment type of the Dynamics CRM instance. 'Online' - for Dynamics CRM Online and 'OnPremisesWithIfd' for Dynamics CRM on-premises with Ifd. Type: - string (or Expression with resultType string). Possible values include: "Online", - "OnPremisesWithIfd". - :type deployment_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsDeploymentType - :param host_name: The host name of the on-premises Dynamics CRM server. The property is - required for on-prem and not allowed for online. Type: string (or Expression with resultType - string). - :type host_name: any - :param port: The port of on-premises Dynamics CRM server. The property is required for on-prem - and not allowed for online. Default is 443. Type: integer (or Expression with resultType - integer), minimum: 0. - :type port: any - :param service_uri: The URL to the Microsoft Dynamics CRM server. The property is required for - on-line and not allowed for on-prem. Type: string (or Expression with resultType string). - :type service_uri: any - :param organization_name: The organization name of the Dynamics CRM instance. The property is - required for on-prem and required for online when there are more than one Dynamics CRM - instances associated with the user. Type: string (or Expression with resultType string). - :type organization_name: any - :param authentication_type: Required. The authentication type to connect to Dynamics CRM - server. 'Office365' for online scenario, 'Ifd' for on-premises with Ifd scenario, - 'AADServicePrincipal' for Server-To-Server authentication in online scenario. Type: string (or - Expression with resultType string). Possible values include: "Office365", "Ifd", - "AADServicePrincipal". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsAuthenticationType - :param username: User name to access the Dynamics CRM instance. Type: string (or Expression - with resultType string). - :type username: any - :param password: Password to access the Dynamics CRM instance. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param service_principal_id: The client ID of the application in Azure Active Directory used - for Server-To-Server authentication. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_credential_type: A string from ServicePrincipalCredentialEnum or an - expression. - :type service_principal_credential_type: any - :param service_principal_credential: The credential of the service principal object in Azure - Active Directory. If servicePrincipalCredentialType is 'ServicePrincipalKey', - servicePrincipalCredential can be SecureString or AzureKeyVaultSecretReference. If - servicePrincipalCredentialType is 'ServicePrincipalCert', servicePrincipalCredential can only - be AzureKeyVaultSecretReference. - :type service_principal_credential: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'deployment_type': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'str'}, - 'host_name': {'key': 'typeProperties.hostName', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'service_uri': {'key': 'typeProperties.serviceUri', 'type': 'object'}, - 'organization_name': {'key': 'typeProperties.organizationName', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_credential_type': {'key': 'typeProperties.servicePrincipalCredentialType', 'type': 'object'}, - 'service_principal_credential': {'key': 'typeProperties.servicePrincipalCredential', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - deployment_type: Union[str, "DynamicsDeploymentType"], - authentication_type: Union[str, "DynamicsAuthenticationType"], - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - host_name: Optional[Any] = None, - port: Optional[Any] = None, - service_uri: Optional[Any] = None, - organization_name: Optional[Any] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - service_principal_id: Optional[Any] = None, - service_principal_credential_type: Optional[Any] = None, - service_principal_credential: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(DynamicsCrmLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'DynamicsCrm' # type: str - self.deployment_type = deployment_type - self.host_name = host_name - self.port = port - self.service_uri = service_uri - self.organization_name = organization_name - self.authentication_type = authentication_type - self.username = username - self.password = password - self.service_principal_id = service_principal_id - self.service_principal_credential_type = service_principal_credential_type - self.service_principal_credential = service_principal_credential - self.encrypted_credential = encrypted_credential - - -class DynamicsCrmSink(CopySink): - """A copy activity Dynamics CRM sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param write_behavior: Required. The write behavior for the operation. Possible values include: - "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsSinkWriteBehavior - :param ignore_null_values: The flag indicating whether to ignore null values from input dataset - (except key fields) during write operation. Default is false. Type: boolean (or Expression with - resultType boolean). - :type ignore_null_values: any - :param alternate_key_name: The logical name of the alternate key which will be used when - upserting records. Type: string (or Expression with resultType string). - :type alternate_key_name: any - """ - - _validation = { - 'type': {'required': True}, - 'write_behavior': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, - 'ignore_null_values': {'key': 'ignoreNullValues', 'type': 'object'}, - 'alternate_key_name': {'key': 'alternateKeyName', 'type': 'object'}, - } - - def __init__( - self, - *, - write_behavior: Union[str, "DynamicsSinkWriteBehavior"], - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - ignore_null_values: Optional[Any] = None, - alternate_key_name: Optional[Any] = None, - **kwargs - ): - super(DynamicsCrmSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'DynamicsCrmSink' # type: str - self.write_behavior = write_behavior - self.ignore_null_values = ignore_null_values - self.alternate_key_name = alternate_key_name - - -class DynamicsCrmSource(CopySource): - """A copy activity Dynamics CRM source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: FetchXML is a proprietary query language that is used in Microsoft Dynamics CRM - (online & on-premises). Type: string (or Expression with resultType string). - :type query: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query: Optional[Any] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(DynamicsCrmSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'DynamicsCrmSource' # type: str - self.query = query - self.additional_columns = additional_columns - - -class DynamicsEntityDataset(Dataset): - """The Dynamics entity dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param entity_name: The logical name of the entity. Type: string (or Expression with resultType - string). - :type entity_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'entity_name': {'key': 'typeProperties.entityName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - entity_name: Optional[Any] = None, - **kwargs - ): - super(DynamicsEntityDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'DynamicsEntity' # type: str - self.entity_name = entity_name - - -class DynamicsLinkedService(LinkedService): - """Dynamics linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param deployment_type: Required. The deployment type of the Dynamics instance. 'Online' for - Dynamics Online and 'OnPremisesWithIfd' for Dynamics on-premises with Ifd. Type: string (or - Expression with resultType string). Possible values include: "Online", "OnPremisesWithIfd". - :type deployment_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsDeploymentType - :param host_name: The host name of the on-premises Dynamics server. The property is required - for on-prem and not allowed for online. Type: string (or Expression with resultType string). - :type host_name: any - :param port: The port of on-premises Dynamics server. The property is required for on-prem and - not allowed for online. Default is 443. Type: integer (or Expression with resultType integer), - minimum: 0. - :type port: any - :param service_uri: The URL to the Microsoft Dynamics server. The property is required for - on-line and not allowed for on-prem. Type: string (or Expression with resultType string). - :type service_uri: any - :param organization_name: The organization name of the Dynamics instance. The property is - required for on-prem and required for online when there are more than one Dynamics instances - associated with the user. Type: string (or Expression with resultType string). - :type organization_name: any - :param authentication_type: Required. The authentication type to connect to Dynamics server. - 'Office365' for online scenario, 'Ifd' for on-premises with Ifd scenario, 'AADServicePrincipal' - for Server-To-Server authentication in online scenario. Type: string (or Expression with - resultType string). Possible values include: "Office365", "Ifd", "AADServicePrincipal". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsAuthenticationType - :param username: User name to access the Dynamics instance. Type: string (or Expression with - resultType string). - :type username: any - :param password: Password to access the Dynamics instance. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param service_principal_id: The client ID of the application in Azure Active Directory used - for Server-To-Server authentication. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_credential_type: The service principal credential type to use in - Server-To-Server authentication. 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' - for certificate. Type: string (or Expression with resultType string). Possible values include: - "ServicePrincipalKey", "ServicePrincipalCert". - :type service_principal_credential_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsServicePrincipalCredentialType - :param service_principal_credential: The credential of the service principal object in Azure - Active Directory. If servicePrincipalCredentialType is 'ServicePrincipalKey', - servicePrincipalCredential can be SecureString or AzureKeyVaultSecretReference. If - servicePrincipalCredentialType is 'ServicePrincipalCert', servicePrincipalCredential can only - be AzureKeyVaultSecretReference. - :type service_principal_credential: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'deployment_type': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'deployment_type': {'key': 'typeProperties.deploymentType', 'type': 'str'}, - 'host_name': {'key': 'typeProperties.hostName', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'service_uri': {'key': 'typeProperties.serviceUri', 'type': 'object'}, - 'organization_name': {'key': 'typeProperties.organizationName', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_credential_type': {'key': 'typeProperties.servicePrincipalCredentialType', 'type': 'str'}, - 'service_principal_credential': {'key': 'typeProperties.servicePrincipalCredential', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - deployment_type: Union[str, "DynamicsDeploymentType"], - authentication_type: Union[str, "DynamicsAuthenticationType"], - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - host_name: Optional[Any] = None, - port: Optional[Any] = None, - service_uri: Optional[Any] = None, - organization_name: Optional[Any] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - service_principal_id: Optional[Any] = None, - service_principal_credential_type: Optional[Union[str, "DynamicsServicePrincipalCredentialType"]] = None, - service_principal_credential: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(DynamicsLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Dynamics' # type: str - self.deployment_type = deployment_type - self.host_name = host_name - self.port = port - self.service_uri = service_uri - self.organization_name = organization_name - self.authentication_type = authentication_type - self.username = username - self.password = password - self.service_principal_id = service_principal_id - self.service_principal_credential_type = service_principal_credential_type - self.service_principal_credential = service_principal_credential - self.encrypted_credential = encrypted_credential - - -class DynamicsSink(CopySink): - """A copy activity Dynamics sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param write_behavior: Required. The write behavior for the operation. Possible values include: - "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2020_12_01.models.DynamicsSinkWriteBehavior - :param ignore_null_values: The flag indicating whether ignore null values from input dataset - (except key fields) during write operation. Default is false. Type: boolean (or Expression with - resultType boolean). - :type ignore_null_values: any - :param alternate_key_name: The logical name of the alternate key which will be used when - upserting records. Type: string (or Expression with resultType string). - :type alternate_key_name: any - """ - - _validation = { - 'type': {'required': True}, - 'write_behavior': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, - 'ignore_null_values': {'key': 'ignoreNullValues', 'type': 'object'}, - 'alternate_key_name': {'key': 'alternateKeyName', 'type': 'object'}, - } - - def __init__( - self, - *, - write_behavior: Union[str, "DynamicsSinkWriteBehavior"], - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - ignore_null_values: Optional[Any] = None, - alternate_key_name: Optional[Any] = None, - **kwargs - ): - super(DynamicsSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'DynamicsSink' # type: str - self.write_behavior = write_behavior - self.ignore_null_values = ignore_null_values - self.alternate_key_name = alternate_key_name - - -class DynamicsSource(CopySource): - """A copy activity Dynamics source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: FetchXML is a proprietary query language that is used in Microsoft Dynamics - (online & on-premises). Type: string (or Expression with resultType string). - :type query: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query: Optional[Any] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(DynamicsSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'DynamicsSource' # type: str - self.query = query - self.additional_columns = additional_columns - - -class EloquaLinkedService(LinkedService): - """Eloqua server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param endpoint: Required. The endpoint of the Eloqua server. (i.e. eloqua.example.com). - :type endpoint: any - :param username: Required. The site name and user name of your Eloqua account in the form: - sitename/username. (i.e. Eloqua/Alice). - :type username: any - :param password: The password corresponding to the user name. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'endpoint': {'required': True}, - 'username': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - endpoint: Any, - username: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - password: Optional["SecretBase"] = None, - use_encrypted_endpoints: Optional[Any] = None, - use_host_verification: Optional[Any] = None, - use_peer_verification: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(EloquaLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Eloqua' # type: str - self.endpoint = endpoint - self.username = username - self.password = password - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential - - -class EloquaObjectDataset(Dataset): - """Eloqua server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(EloquaObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'EloquaObject' # type: str - self.table_name = table_name - - -class EloquaSource(TabularSource): - """A copy activity Eloqua server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(EloquaSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'EloquaSource' # type: str - self.query = query - - -class EncryptionDetails(msrest.serialization.Model): - """Details of the encryption associated with the workspace. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar double_encryption_enabled: Double Encryption enabled. - :vartype double_encryption_enabled: bool - :param cmk: Customer Managed Key Details. - :type cmk: ~azure.synapse.artifacts.v2020_12_01.models.CustomerManagedKeyDetails - """ - - _validation = { - 'double_encryption_enabled': {'readonly': True}, - } - - _attribute_map = { - 'double_encryption_enabled': {'key': 'doubleEncryptionEnabled', 'type': 'bool'}, - 'cmk': {'key': 'cmk', 'type': 'CustomerManagedKeyDetails'}, - } - - def __init__( - self, - *, - cmk: Optional["CustomerManagedKeyDetails"] = None, - **kwargs - ): - super(EncryptionDetails, self).__init__(**kwargs) - self.double_encryption_enabled = None - self.cmk = cmk - - -class EntityReference(msrest.serialization.Model): - """The entity reference. - - :param type: The type of this referenced entity. Possible values include: - "IntegrationRuntimeReference", "LinkedServiceReference". - :type type: str or - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeEntityReferenceType - :param reference_name: The name of this referenced entity. - :type reference_name: str - """ - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - } - - def __init__( - self, - *, - type: Optional[Union[str, "IntegrationRuntimeEntityReferenceType"]] = None, - reference_name: Optional[str] = None, - **kwargs - ): - super(EntityReference, self).__init__(**kwargs) - self.type = type - self.reference_name = reference_name - - -class ErrorAdditionalInfo(msrest.serialization.Model): - """The resource management error additional info. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar type: The additional info type. - :vartype type: str - :ivar info: The additional info. - :vartype info: any - """ - - _validation = { - 'type': {'readonly': True}, - 'info': {'readonly': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'info': {'key': 'info', 'type': 'object'}, - } - - def __init__( - self, - **kwargs - ): - super(ErrorAdditionalInfo, self).__init__(**kwargs) - self.type = None - self.info = None - - -class ErrorContract(msrest.serialization.Model): - """Contains details when the response code indicates an error. - - :param error: The error details. - :type error: ~azure.synapse.artifacts.v2020_12_01.models.ErrorResponse - """ - - _attribute_map = { - 'error': {'key': 'error', 'type': 'ErrorResponse'}, - } - - def __init__( - self, - *, - error: Optional["ErrorResponse"] = None, - **kwargs - ): - super(ErrorContract, self).__init__(**kwargs) - self.error = error - - -class ErrorResponse(msrest.serialization.Model): - """Common error response for all Azure Resource Manager APIs to return error details for failed operations. (This also follows the OData error response format.). - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar code: The error code. - :vartype code: str - :ivar message: The error message. - :vartype message: str - :ivar target: The error target. - :vartype target: str - :ivar details: The error details. - :vartype details: list[~azure.synapse.artifacts.v2020_12_01.models.ErrorResponse] - :ivar additional_info: The error additional info. - :vartype additional_info: list[~azure.synapse.artifacts.v2020_12_01.models.ErrorAdditionalInfo] - """ - - _validation = { - 'code': {'readonly': True}, - 'message': {'readonly': True}, - 'target': {'readonly': True}, - 'details': {'readonly': True}, - 'additional_info': {'readonly': True}, - } - - _attribute_map = { - 'code': {'key': 'code', 'type': 'str'}, - 'message': {'key': 'message', 'type': 'str'}, - 'target': {'key': 'target', 'type': 'str'}, - 'details': {'key': 'details', 'type': '[ErrorResponse]'}, - 'additional_info': {'key': 'additionalInfo', 'type': '[ErrorAdditionalInfo]'}, - } - - def __init__( - self, - **kwargs - ): - super(ErrorResponse, self).__init__(**kwargs) - self.code = None - self.message = None - self.target = None - self.details = None - self.additional_info = None - - -class EvaluateDataFlowExpressionRequest(msrest.serialization.Model): - """Request body structure for data flow expression preview. - - :param session_id: The ID of data flow debug session. - :type session_id: str - :param data_flow_name: The data flow which contains the debug session. - :type data_flow_name: str - :param stream_name: The output stream name. - :type stream_name: str - :param row_limits: The row limit for preview request. - :type row_limits: int - :param expression: The expression for preview. - :type expression: str - """ - - _attribute_map = { - 'session_id': {'key': 'sessionId', 'type': 'str'}, - 'data_flow_name': {'key': 'dataFlowName', 'type': 'str'}, - 'stream_name': {'key': 'streamName', 'type': 'str'}, - 'row_limits': {'key': 'rowLimits', 'type': 'int'}, - 'expression': {'key': 'expression', 'type': 'str'}, - } - - def __init__( - self, - *, - session_id: Optional[str] = None, - data_flow_name: Optional[str] = None, - stream_name: Optional[str] = None, - row_limits: Optional[int] = None, - expression: Optional[str] = None, - **kwargs - ): - super(EvaluateDataFlowExpressionRequest, self).__init__(**kwargs) - self.session_id = session_id - self.data_flow_name = data_flow_name - self.stream_name = stream_name - self.row_limits = row_limits - self.expression = expression - - -class ExcelDataset(Dataset): - """Excel dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param location: The location of the excel storage. - :type location: ~azure.synapse.artifacts.v2020_12_01.models.DatasetLocation - :param sheet_name: The sheet of excel file. Type: string (or Expression with resultType - string). - :type sheet_name: any - :param range: The partial data of one sheet. Type: string (or Expression with resultType - string). - :type range: any - :param first_row_as_header: When used as input, treat the first row of data as headers. When - used as output,write the headers into the output as the first row of data. The default value is - false. Type: boolean (or Expression with resultType boolean). - :type first_row_as_header: any - :param compression: The data compression method used for the json dataset. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - :param null_value: The null value string. Type: string (or Expression with resultType string). - :type null_value: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'location': {'key': 'typeProperties.location', 'type': 'DatasetLocation'}, - 'sheet_name': {'key': 'typeProperties.sheetName', 'type': 'object'}, - 'range': {'key': 'typeProperties.range', 'type': 'object'}, - 'first_row_as_header': {'key': 'typeProperties.firstRowAsHeader', 'type': 'object'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - 'null_value': {'key': 'typeProperties.nullValue', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - location: Optional["DatasetLocation"] = None, - sheet_name: Optional[Any] = None, - range: Optional[Any] = None, - first_row_as_header: Optional[Any] = None, - compression: Optional["DatasetCompression"] = None, - null_value: Optional[Any] = None, - **kwargs - ): - super(ExcelDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'Excel' # type: str - self.location = location - self.sheet_name = sheet_name - self.range = range - self.first_row_as_header = first_row_as_header - self.compression = compression - self.null_value = null_value - - -class ExcelSource(CopySource): - """A copy activity excel source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Excel store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - store_settings: Optional["StoreReadSettings"] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(ExcelSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'ExcelSource' # type: str - self.store_settings = store_settings - self.additional_columns = additional_columns - - -class ExecuteDataFlowActivity(ExecutionActivity): - """Execute data flow activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param dataflow: Required. Data flow reference. - :type dataflow: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowReference - :param staging: Staging info for execute data flow activity. - :type staging: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowStagingInfo - :param integration_runtime: The integration runtime reference. - :type integration_runtime: - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param compute: Compute properties for data flow activity. - :type compute: - ~azure.synapse.artifacts.v2020_12_01.models.ExecuteDataFlowActivityTypePropertiesCompute - :param trace_level: Trace level setting used for data flow monitoring output. Supported values - are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string). - :type trace_level: any - :param continue_on_error: Continue on error setting used for data flow execution. Enables - processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean). - :type continue_on_error: any - :param run_concurrently: Concurrent run setting used for data flow execution. Allows sinks with - the same save order to be processed concurrently. Type: boolean (or Expression with resultType - boolean). - :type run_concurrently: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'dataflow': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'dataflow': {'key': 'typeProperties.dataflow', 'type': 'DataFlowReference'}, - 'staging': {'key': 'typeProperties.staging', 'type': 'DataFlowStagingInfo'}, - 'integration_runtime': {'key': 'typeProperties.integrationRuntime', 'type': 'IntegrationRuntimeReference'}, - 'compute': {'key': 'typeProperties.compute', 'type': 'ExecuteDataFlowActivityTypePropertiesCompute'}, - 'trace_level': {'key': 'typeProperties.traceLevel', 'type': 'object'}, - 'continue_on_error': {'key': 'typeProperties.continueOnError', 'type': 'object'}, - 'run_concurrently': {'key': 'typeProperties.runConcurrently', 'type': 'object'}, - } - - def __init__( - self, - *, - name: str, - dataflow: "DataFlowReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - staging: Optional["DataFlowStagingInfo"] = None, - integration_runtime: Optional["IntegrationRuntimeReference"] = None, - compute: Optional["ExecuteDataFlowActivityTypePropertiesCompute"] = None, - trace_level: Optional[Any] = None, - continue_on_error: Optional[Any] = None, - run_concurrently: Optional[Any] = None, - **kwargs - ): - super(ExecuteDataFlowActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'ExecuteDataFlow' # type: str - self.dataflow = dataflow - self.staging = staging - self.integration_runtime = integration_runtime - self.compute = compute - self.trace_level = trace_level - self.continue_on_error = continue_on_error - self.run_concurrently = run_concurrently - - -class ExecuteDataFlowActivityTypePropertiesCompute(msrest.serialization.Model): - """Compute properties for data flow activity. - - :param compute_type: Compute type of the cluster which will execute data flow job. Possible - values include: "General", "MemoryOptimized", "ComputeOptimized". - :type compute_type: str or ~azure.synapse.artifacts.v2020_12_01.models.DataFlowComputeType - :param core_count: Core count of the cluster which will execute data flow job. Supported values - are: 8, 16, 32, 48, 80, 144 and 272. - :type core_count: int - """ - - _attribute_map = { - 'compute_type': {'key': 'computeType', 'type': 'str'}, - 'core_count': {'key': 'coreCount', 'type': 'int'}, - } - - def __init__( - self, - *, - compute_type: Optional[Union[str, "DataFlowComputeType"]] = None, - core_count: Optional[int] = None, - **kwargs - ): - super(ExecuteDataFlowActivityTypePropertiesCompute, self).__init__(**kwargs) - self.compute_type = compute_type - self.core_count = core_count - - -class ExecutePipelineActivity(ControlActivity): - """Execute pipeline activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param pipeline: Required. Pipeline reference. - :type pipeline: ~azure.synapse.artifacts.v2020_12_01.models.PipelineReference - :param parameters: Pipeline parameters. - :type parameters: dict[str, any] - :param wait_on_completion: Defines whether activity execution will wait for the dependent - pipeline execution to finish. Default is false. - :type wait_on_completion: bool - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'pipeline': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'pipeline': {'key': 'typeProperties.pipeline', 'type': 'PipelineReference'}, - 'parameters': {'key': 'typeProperties.parameters', 'type': '{object}'}, - 'wait_on_completion': {'key': 'typeProperties.waitOnCompletion', 'type': 'bool'}, - } - - def __init__( - self, - *, - name: str, - pipeline: "PipelineReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - parameters: Optional[Dict[str, Any]] = None, - wait_on_completion: Optional[bool] = None, - **kwargs - ): - super(ExecutePipelineActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) - self.type = 'ExecutePipeline' # type: str - self.pipeline = pipeline - self.parameters = parameters - self.wait_on_completion = wait_on_completion - - -class ExecuteSSISPackageActivity(ExecutionActivity): - """Execute SSIS package activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param package_location: Required. SSIS package location. - :type package_location: ~azure.synapse.artifacts.v2020_12_01.models.SSISPackageLocation - :param runtime: Specifies the runtime to execute SSIS package. The value should be "x86" or - "x64". Type: string (or Expression with resultType string). - :type runtime: any - :param logging_level: The logging level of SSIS package execution. Type: string (or Expression - with resultType string). - :type logging_level: any - :param environment_path: The environment path to execute the SSIS package. Type: string (or - Expression with resultType string). - :type environment_path: any - :param execution_credential: The package execution credential. - :type execution_credential: ~azure.synapse.artifacts.v2020_12_01.models.SSISExecutionCredential - :param connect_via: Required. The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param project_parameters: The project level parameters to execute the SSIS package. - :type project_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.SSISExecutionParameter] - :param package_parameters: The package level parameters to execute the SSIS package. - :type package_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.SSISExecutionParameter] - :param project_connection_managers: The project level connection managers to execute the SSIS - package. - :type project_connection_managers: dict[str, dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.SSISExecutionParameter]] - :param package_connection_managers: The package level connection managers to execute the SSIS - package. - :type package_connection_managers: dict[str, dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.SSISExecutionParameter]] - :param property_overrides: The property overrides to execute the SSIS package. - :type property_overrides: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.SSISPropertyOverride] - :param log_location: SSIS package execution log location. - :type log_location: ~azure.synapse.artifacts.v2020_12_01.models.SSISLogLocation - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'package_location': {'required': True}, - 'connect_via': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'package_location': {'key': 'typeProperties.packageLocation', 'type': 'SSISPackageLocation'}, - 'runtime': {'key': 'typeProperties.runtime', 'type': 'object'}, - 'logging_level': {'key': 'typeProperties.loggingLevel', 'type': 'object'}, - 'environment_path': {'key': 'typeProperties.environmentPath', 'type': 'object'}, - 'execution_credential': {'key': 'typeProperties.executionCredential', 'type': 'SSISExecutionCredential'}, - 'connect_via': {'key': 'typeProperties.connectVia', 'type': 'IntegrationRuntimeReference'}, - 'project_parameters': {'key': 'typeProperties.projectParameters', 'type': '{SSISExecutionParameter}'}, - 'package_parameters': {'key': 'typeProperties.packageParameters', 'type': '{SSISExecutionParameter}'}, - 'project_connection_managers': {'key': 'typeProperties.projectConnectionManagers', 'type': '{{SSISExecutionParameter}}'}, - 'package_connection_managers': {'key': 'typeProperties.packageConnectionManagers', 'type': '{{SSISExecutionParameter}}'}, - 'property_overrides': {'key': 'typeProperties.propertyOverrides', 'type': '{SSISPropertyOverride}'}, - 'log_location': {'key': 'typeProperties.logLocation', 'type': 'SSISLogLocation'}, - } - - def __init__( - self, - *, - name: str, - package_location: "SSISPackageLocation", - connect_via: "IntegrationRuntimeReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - runtime: Optional[Any] = None, - logging_level: Optional[Any] = None, - environment_path: Optional[Any] = None, - execution_credential: Optional["SSISExecutionCredential"] = None, - project_parameters: Optional[Dict[str, "SSISExecutionParameter"]] = None, - package_parameters: Optional[Dict[str, "SSISExecutionParameter"]] = None, - project_connection_managers: Optional[Dict[str, Dict[str, "SSISExecutionParameter"]]] = None, - package_connection_managers: Optional[Dict[str, Dict[str, "SSISExecutionParameter"]]] = None, - property_overrides: Optional[Dict[str, "SSISPropertyOverride"]] = None, - log_location: Optional["SSISLogLocation"] = None, - **kwargs - ): - super(ExecuteSSISPackageActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'ExecuteSSISPackage' # type: str - self.package_location = package_location - self.runtime = runtime - self.logging_level = logging_level - self.environment_path = environment_path - self.execution_credential = execution_credential - self.connect_via = connect_via - self.project_parameters = project_parameters - self.package_parameters = package_parameters - self.project_connection_managers = project_connection_managers - self.package_connection_managers = package_connection_managers - self.property_overrides = property_overrides - self.log_location = log_location - - -class ExposureControlRequest(msrest.serialization.Model): - """The exposure control request. - - :param feature_name: The feature name. - :type feature_name: str - :param feature_type: The feature type. - :type feature_type: str - """ - - _attribute_map = { - 'feature_name': {'key': 'featureName', 'type': 'str'}, - 'feature_type': {'key': 'featureType', 'type': 'str'}, - } - - def __init__( - self, - *, - feature_name: Optional[str] = None, - feature_type: Optional[str] = None, - **kwargs - ): - super(ExposureControlRequest, self).__init__(**kwargs) - self.feature_name = feature_name - self.feature_type = feature_type - - -class ExposureControlResponse(msrest.serialization.Model): - """The exposure control response. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar feature_name: The feature name. - :vartype feature_name: str - :ivar value: The feature value. - :vartype value: str - """ - - _validation = { - 'feature_name': {'readonly': True}, - 'value': {'readonly': True}, - } - - _attribute_map = { - 'feature_name': {'key': 'featureName', 'type': 'str'}, - 'value': {'key': 'value', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(ExposureControlResponse, self).__init__(**kwargs) - self.feature_name = None - self.value = None - - -class Expression(msrest.serialization.Model): - """Azure Synapse expression definition. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Expression type. Possible values include: "Expression". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.ExpressionType - :param value: Required. Expression value. - :type value: str - """ - - _validation = { - 'type': {'required': True}, - 'value': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'value': {'key': 'value', 'type': 'str'}, - } - - def __init__( - self, - *, - type: Union[str, "ExpressionType"], - value: str, - **kwargs - ): - super(Expression, self).__init__(**kwargs) - self.type = type - self.value = value - - -class FileServerLinkedService(LinkedService): - """File system linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. Host name of the server. Type: string (or Expression with resultType - string). - :type host: any - :param user_id: User ID to logon the server. Type: string (or Expression with resultType - string). - :type user_id: any - :param password: Password to logon the server. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'user_id': {'key': 'typeProperties.userId', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - host: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - user_id: Optional[Any] = None, - password: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(FileServerLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'FileServer' # type: str - self.host = host - self.user_id = user_id - self.password = password - self.encrypted_credential = encrypted_credential - - -class FileServerLocation(DatasetLocation): - """The location of file server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - folder_path: Optional[Any] = None, - file_name: Optional[Any] = None, - **kwargs - ): - super(FileServerLocation, self).__init__(additional_properties=additional_properties, folder_path=folder_path, file_name=file_name, **kwargs) - self.type = 'FileServerLocation' # type: str - - -class FileServerReadSettings(StoreReadSettings): - """File server read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: FileServer wildcardFolderPath. Type: string (or Expression with - resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: FileServer wildcardFileName. Type: string (or Expression with - resultType string). - :type wildcard_file_name: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - :param file_filter: Specify a filter to be used to select a subset of files in the folderPath - rather than all files. Type: string (or Expression with resultType string). - :type file_filter: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - 'modified_datetime_start': {'key': 'modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'modifiedDatetimeEnd', 'type': 'object'}, - 'file_filter': {'key': 'fileFilter', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - recursive: Optional[Any] = None, - wildcard_folder_path: Optional[Any] = None, - wildcard_file_name: Optional[Any] = None, - file_list_path: Optional[Any] = None, - enable_partition_discovery: Optional[bool] = None, - partition_root_path: Optional[Any] = None, - delete_files_after_completion: Optional[Any] = None, - modified_datetime_start: Optional[Any] = None, - modified_datetime_end: Optional[Any] = None, - file_filter: Optional[Any] = None, - **kwargs - ): - super(FileServerReadSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'FileServerReadSettings' # type: str - self.recursive = recursive - self.wildcard_folder_path = wildcard_folder_path - self.wildcard_file_name = wildcard_file_name - self.file_list_path = file_list_path - self.enable_partition_discovery = enable_partition_discovery - self.partition_root_path = partition_root_path - self.delete_files_after_completion = delete_files_after_completion - self.modified_datetime_start = modified_datetime_start - self.modified_datetime_end = modified_datetime_end - self.file_filter = file_filter - - -class FileServerWriteSettings(StoreWriteSettings): - """File server write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - copy_behavior: Optional[Any] = None, - **kwargs - ): - super(FileServerWriteSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) - self.type = 'FileServerWriteSettings' # type: str - - -class FileShareDataset(Dataset): - """An on-premises file system dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param folder_path: The path of the on-premises file system. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: The name of the on-premises file system. Type: string (or Expression with - resultType string). - :type file_name: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - :param format: The format of the files. - :type format: ~azure.synapse.artifacts.v2020_12_01.models.DatasetStorageFormat - :param file_filter: Specify a filter to be used to select a subset of files in the folderPath - rather than all files. Type: string (or Expression with resultType string). - :type file_filter: any - :param compression: The data compression method used for the file system. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'object'}, - 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, - 'modified_datetime_start': {'key': 'typeProperties.modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'typeProperties.modifiedDatetimeEnd', 'type': 'object'}, - 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, - 'file_filter': {'key': 'typeProperties.fileFilter', 'type': 'object'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - folder_path: Optional[Any] = None, - file_name: Optional[Any] = None, - modified_datetime_start: Optional[Any] = None, - modified_datetime_end: Optional[Any] = None, - format: Optional["DatasetStorageFormat"] = None, - file_filter: Optional[Any] = None, - compression: Optional["DatasetCompression"] = None, - **kwargs - ): - super(FileShareDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'FileShare' # type: str - self.folder_path = folder_path - self.file_name = file_name - self.modified_datetime_start = modified_datetime_start - self.modified_datetime_end = modified_datetime_end - self.format = format - self.file_filter = file_filter - self.compression = compression - - -class FileSystemSink(CopySink): - """A copy activity file system sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - copy_behavior: Optional[Any] = None, - **kwargs - ): - super(FileSystemSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'FileSystemSink' # type: str - self.copy_behavior = copy_behavior - - -class FileSystemSource(CopySource): - """A copy activity file system source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - recursive: Optional[Any] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(FileSystemSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'FileSystemSource' # type: str - self.recursive = recursive - self.additional_columns = additional_columns - - -class FilterActivity(ControlActivity): - """Filter and return results from input array based on the conditions. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param items: Required. Input array on which filter should be applied. - :type items: ~azure.synapse.artifacts.v2020_12_01.models.Expression - :param condition: Required. Condition to be used for filtering the input. - :type condition: ~azure.synapse.artifacts.v2020_12_01.models.Expression - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'items': {'required': True}, - 'condition': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'items': {'key': 'typeProperties.items', 'type': 'Expression'}, - 'condition': {'key': 'typeProperties.condition', 'type': 'Expression'}, - } - - def __init__( - self, - *, - name: str, - items: "Expression", - condition: "Expression", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - **kwargs - ): - super(FilterActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) - self.type = 'Filter' # type: str - self.items = items - self.condition = condition - - -class ForEachActivity(ControlActivity): - """This activity is used for iterating over a collection and execute given activities. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param is_sequential: Should the loop be executed in sequence or in parallel (max 50). - :type is_sequential: bool - :param batch_count: Batch count to be used for controlling the number of parallel execution - (when isSequential is set to false). - :type batch_count: int - :param items: Required. Collection to iterate. - :type items: ~azure.synapse.artifacts.v2020_12_01.models.Expression - :param activities: Required. List of activities to execute . - :type activities: list[~azure.synapse.artifacts.v2020_12_01.models.Activity] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'batch_count': {'maximum': 50}, - 'items': {'required': True}, - 'activities': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'is_sequential': {'key': 'typeProperties.isSequential', 'type': 'bool'}, - 'batch_count': {'key': 'typeProperties.batchCount', 'type': 'int'}, - 'items': {'key': 'typeProperties.items', 'type': 'Expression'}, - 'activities': {'key': 'typeProperties.activities', 'type': '[Activity]'}, - } - - def __init__( - self, - *, - name: str, - items: "Expression", - activities: List["Activity"], - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - is_sequential: Optional[bool] = None, - batch_count: Optional[int] = None, - **kwargs - ): - super(ForEachActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) - self.type = 'ForEach' # type: str - self.is_sequential = is_sequential - self.batch_count = batch_count - self.items = items - self.activities = activities - - -class FtpReadSettings(StoreReadSettings): - """Ftp read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: Ftp wildcardFolderPath. Type: string (or Expression with - resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: Ftp wildcardFileName. Type: string (or Expression with resultType - string). - :type wildcard_file_name: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param use_binary_transfer: Specify whether to use binary transfer mode for FTP stores. - :type use_binary_transfer: bool - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'use_binary_transfer': {'key': 'useBinaryTransfer', 'type': 'bool'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - recursive: Optional[Any] = None, - wildcard_folder_path: Optional[Any] = None, - wildcard_file_name: Optional[Any] = None, - enable_partition_discovery: Optional[bool] = None, - partition_root_path: Optional[Any] = None, - delete_files_after_completion: Optional[Any] = None, - file_list_path: Optional[Any] = None, - use_binary_transfer: Optional[bool] = None, - **kwargs - ): - super(FtpReadSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'FtpReadSettings' # type: str - self.recursive = recursive - self.wildcard_folder_path = wildcard_folder_path - self.wildcard_file_name = wildcard_file_name - self.enable_partition_discovery = enable_partition_discovery - self.partition_root_path = partition_root_path - self.delete_files_after_completion = delete_files_after_completion - self.file_list_path = file_list_path - self.use_binary_transfer = use_binary_transfer - - -class FtpServerLinkedService(LinkedService): - """A FTP server Linked Service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. Host name of the FTP server. Type: string (or Expression with resultType - string). - :type host: any - :param port: The TCP port number that the FTP server uses to listen for client connections. - Default value is 21. Type: integer (or Expression with resultType integer), minimum: 0. - :type port: any - :param authentication_type: The authentication type to be used to connect to the FTP server. - Possible values include: "Basic", "Anonymous". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.FtpAuthenticationType - :param user_name: Username to logon the FTP server. Type: string (or Expression with resultType - string). - :type user_name: any - :param password: Password to logon the FTP server. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - :param enable_ssl: If true, connect to the FTP server over SSL/TLS channel. Default value is - true. Type: boolean (or Expression with resultType boolean). - :type enable_ssl: any - :param enable_server_certificate_validation: If true, validate the FTP server SSL certificate - when connect over SSL/TLS channel. Default value is true. Type: boolean (or Expression with - resultType boolean). - :type enable_server_certificate_validation: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, - 'enable_server_certificate_validation': {'key': 'typeProperties.enableServerCertificateValidation', 'type': 'object'}, - } - - def __init__( - self, - *, - host: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - port: Optional[Any] = None, - authentication_type: Optional[Union[str, "FtpAuthenticationType"]] = None, - user_name: Optional[Any] = None, - password: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - enable_ssl: Optional[Any] = None, - enable_server_certificate_validation: Optional[Any] = None, - **kwargs - ): - super(FtpServerLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'FtpServer' # type: str - self.host = host - self.port = port - self.authentication_type = authentication_type - self.user_name = user_name - self.password = password - self.encrypted_credential = encrypted_credential - self.enable_ssl = enable_ssl - self.enable_server_certificate_validation = enable_server_certificate_validation - - -class FtpServerLocation(DatasetLocation): - """The location of ftp server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - folder_path: Optional[Any] = None, - file_name: Optional[Any] = None, - **kwargs - ): - super(FtpServerLocation, self).__init__(additional_properties=additional_properties, folder_path=folder_path, file_name=file_name, **kwargs) - self.type = 'FtpServerLocation' # type: str - - -class GetMetadataActivity(ExecutionActivity): - """Activity to get metadata of dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param dataset: Required. GetMetadata activity dataset reference. - :type dataset: ~azure.synapse.artifacts.v2020_12_01.models.DatasetReference - :param field_list: Fields of metadata to get from dataset. - :type field_list: list[any] - :param store_settings: GetMetadata activity store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - :param format_settings: GetMetadata activity format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.FormatReadSettings - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'dataset': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'dataset': {'key': 'typeProperties.dataset', 'type': 'DatasetReference'}, - 'field_list': {'key': 'typeProperties.fieldList', 'type': '[object]'}, - 'store_settings': {'key': 'typeProperties.storeSettings', 'type': 'StoreReadSettings'}, - 'format_settings': {'key': 'typeProperties.formatSettings', 'type': 'FormatReadSettings'}, - } - - def __init__( - self, - *, - name: str, - dataset: "DatasetReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - field_list: Optional[List[Any]] = None, - store_settings: Optional["StoreReadSettings"] = None, - format_settings: Optional["FormatReadSettings"] = None, - **kwargs - ): - super(GetMetadataActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'GetMetadata' # type: str - self.dataset = dataset - self.field_list = field_list - self.store_settings = store_settings - self.format_settings = format_settings - - -class GetSsisObjectMetadataRequest(msrest.serialization.Model): - """The request payload of get SSIS object metadata. - - :param metadata_path: Metadata path. - :type metadata_path: str - """ - - _attribute_map = { - 'metadata_path': {'key': 'metadataPath', 'type': 'str'}, - } - - def __init__( - self, - *, - metadata_path: Optional[str] = None, - **kwargs - ): - super(GetSsisObjectMetadataRequest, self).__init__(**kwargs) - self.metadata_path = metadata_path - - -class GitHubAccessTokenRequest(msrest.serialization.Model): - """GitHubAccessTokenRequest. - - All required parameters must be populated in order to send to Azure. - - :param git_hub_client_id: Required. The GitHub Client Id. - :type git_hub_client_id: str - :param git_hub_access_code: Required. The GitHub Access code. - :type git_hub_access_code: str - :param git_hub_access_token_base_url: Required. The GitHub access token base URL. - :type git_hub_access_token_base_url: str - """ - - _validation = { - 'git_hub_client_id': {'required': True}, - 'git_hub_access_code': {'required': True}, - 'git_hub_access_token_base_url': {'required': True}, - } - - _attribute_map = { - 'git_hub_client_id': {'key': 'gitHubClientId', 'type': 'str'}, - 'git_hub_access_code': {'key': 'gitHubAccessCode', 'type': 'str'}, - 'git_hub_access_token_base_url': {'key': 'gitHubAccessTokenBaseUrl', 'type': 'str'}, - } - - def __init__( - self, - *, - git_hub_client_id: str, - git_hub_access_code: str, - git_hub_access_token_base_url: str, - **kwargs - ): - super(GitHubAccessTokenRequest, self).__init__(**kwargs) - self.git_hub_client_id = git_hub_client_id - self.git_hub_access_code = git_hub_access_code - self.git_hub_access_token_base_url = git_hub_access_token_base_url - - -class GitHubAccessTokenResponse(msrest.serialization.Model): - """GitHubAccessTokenResponse. - - :param git_hub_access_token: - :type git_hub_access_token: str - """ - - _attribute_map = { - 'git_hub_access_token': {'key': 'gitHubAccessToken', 'type': 'str'}, - } - - def __init__( - self, - *, - git_hub_access_token: Optional[str] = None, - **kwargs - ): - super(GitHubAccessTokenResponse, self).__init__(**kwargs) - self.git_hub_access_token = git_hub_access_token - - -class GitHubClientSecret(msrest.serialization.Model): - """Client secret information for factory's bring your own app repository configuration. - - :param byoa_secret_akv_url: Bring your own app client secret AKV URL. - :type byoa_secret_akv_url: str - :param byoa_secret_name: Bring your own app client secret name in AKV. - :type byoa_secret_name: str - """ - - _attribute_map = { - 'byoa_secret_akv_url': {'key': 'byoaSecretAkvUrl', 'type': 'str'}, - 'byoa_secret_name': {'key': 'byoaSecretName', 'type': 'str'}, - } - - def __init__( - self, - *, - byoa_secret_akv_url: Optional[str] = None, - byoa_secret_name: Optional[str] = None, - **kwargs - ): - super(GitHubClientSecret, self).__init__(**kwargs) - self.byoa_secret_akv_url = byoa_secret_akv_url - self.byoa_secret_name = byoa_secret_name - - -class GoogleAdWordsLinkedService(LinkedService): - """Google AdWords service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param client_customer_id: Required. The Client customer ID of the AdWords account that you - want to fetch report data for. - :type client_customer_id: any - :param developer_token: Required. The developer token associated with the manager account that - you use to grant access to the AdWords API. - :type developer_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param authentication_type: Required. The OAuth 2.0 authentication mechanism used for - authentication. ServiceAuthentication can only be used on self-hosted IR. Possible values - include: "ServiceAuthentication", "UserAuthentication". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.GoogleAdWordsAuthenticationType - :param refresh_token: The refresh token obtained from Google for authorizing access to AdWords - for UserAuthentication. - :type refresh_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param client_id: The client id of the google application used to acquire the refresh token. - Type: string (or Expression with resultType string). - :type client_id: any - :param client_secret: The client secret of the google application used to acquire the refresh - token. - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param email: The service account email ID that is used for ServiceAuthentication and can only - be used on self-hosted IR. - :type email: any - :param key_file_path: The full path to the .p12 key file that is used to authenticate the - service account email address and can only be used on self-hosted IR. - :type key_file_path: any - :param trusted_cert_path: The full path of the .pem file containing trusted CA certificates for - verifying the server when connecting over SSL. This property can only be set when using SSL on - self-hosted IR. The default value is the cacerts.pem file installed with the IR. - :type trusted_cert_path: any - :param use_system_trust_store: Specifies whether to use a CA certificate from the system trust - store or from a specified PEM file. The default value is false. - :type use_system_trust_store: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'client_customer_id': {'required': True}, - 'developer_token': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'client_customer_id': {'key': 'typeProperties.clientCustomerID', 'type': 'object'}, - 'developer_token': {'key': 'typeProperties.developerToken', 'type': 'SecretBase'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'refresh_token': {'key': 'typeProperties.refreshToken', 'type': 'SecretBase'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'email': {'key': 'typeProperties.email', 'type': 'object'}, - 'key_file_path': {'key': 'typeProperties.keyFilePath', 'type': 'object'}, - 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, - 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - client_customer_id: Any, - developer_token: "SecretBase", - authentication_type: Union[str, "GoogleAdWordsAuthenticationType"], - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - refresh_token: Optional["SecretBase"] = None, - client_id: Optional[Any] = None, - client_secret: Optional["SecretBase"] = None, - email: Optional[Any] = None, - key_file_path: Optional[Any] = None, - trusted_cert_path: Optional[Any] = None, - use_system_trust_store: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(GoogleAdWordsLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'GoogleAdWords' # type: str - self.client_customer_id = client_customer_id - self.developer_token = developer_token - self.authentication_type = authentication_type - self.refresh_token = refresh_token - self.client_id = client_id - self.client_secret = client_secret - self.email = email - self.key_file_path = key_file_path - self.trusted_cert_path = trusted_cert_path - self.use_system_trust_store = use_system_trust_store - self.encrypted_credential = encrypted_credential - - -class GoogleAdWordsObjectDataset(Dataset): - """Google AdWords service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(GoogleAdWordsObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'GoogleAdWordsObject' # type: str - self.table_name = table_name - - -class GoogleAdWordsSource(TabularSource): - """A copy activity Google AdWords service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(GoogleAdWordsSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'GoogleAdWordsSource' # type: str - self.query = query - - -class GoogleBigQueryLinkedService(LinkedService): - """Google BigQuery service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param project: Required. The default BigQuery project to query against. - :type project: any - :param additional_projects: A comma-separated list of public BigQuery projects to access. - :type additional_projects: any - :param request_google_drive_scope: Whether to request access to Google Drive. Allowing Google - Drive access enables support for federated tables that combine BigQuery data with data from - Google Drive. The default value is false. - :type request_google_drive_scope: any - :param authentication_type: Required. The OAuth 2.0 authentication mechanism used for - authentication. ServiceAuthentication can only be used on self-hosted IR. Possible values - include: "ServiceAuthentication", "UserAuthentication". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.GoogleBigQueryAuthenticationType - :param refresh_token: The refresh token obtained from Google for authorizing access to BigQuery - for UserAuthentication. - :type refresh_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param client_id: The client id of the google application used to acquire the refresh token. - Type: string (or Expression with resultType string). - :type client_id: any - :param client_secret: The client secret of the google application used to acquire the refresh - token. - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param email: The service account email ID that is used for ServiceAuthentication and can only - be used on self-hosted IR. - :type email: any - :param key_file_path: The full path to the .p12 key file that is used to authenticate the - service account email address and can only be used on self-hosted IR. - :type key_file_path: any - :param trusted_cert_path: The full path of the .pem file containing trusted CA certificates for - verifying the server when connecting over SSL. This property can only be set when using SSL on - self-hosted IR. The default value is the cacerts.pem file installed with the IR. - :type trusted_cert_path: any - :param use_system_trust_store: Specifies whether to use a CA certificate from the system trust - store or from a specified PEM file. The default value is false. - :type use_system_trust_store: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'project': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'project': {'key': 'typeProperties.project', 'type': 'object'}, - 'additional_projects': {'key': 'typeProperties.additionalProjects', 'type': 'object'}, - 'request_google_drive_scope': {'key': 'typeProperties.requestGoogleDriveScope', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'refresh_token': {'key': 'typeProperties.refreshToken', 'type': 'SecretBase'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'email': {'key': 'typeProperties.email', 'type': 'object'}, - 'key_file_path': {'key': 'typeProperties.keyFilePath', 'type': 'object'}, - 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, - 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - project: Any, - authentication_type: Union[str, "GoogleBigQueryAuthenticationType"], - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - additional_projects: Optional[Any] = None, - request_google_drive_scope: Optional[Any] = None, - refresh_token: Optional["SecretBase"] = None, - client_id: Optional[Any] = None, - client_secret: Optional["SecretBase"] = None, - email: Optional[Any] = None, - key_file_path: Optional[Any] = None, - trusted_cert_path: Optional[Any] = None, - use_system_trust_store: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(GoogleBigQueryLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'GoogleBigQuery' # type: str - self.project = project - self.additional_projects = additional_projects - self.request_google_drive_scope = request_google_drive_scope - self.authentication_type = authentication_type - self.refresh_token = refresh_token - self.client_id = client_id - self.client_secret = client_secret - self.email = email - self.key_file_path = key_file_path - self.trusted_cert_path = trusted_cert_path - self.use_system_trust_store = use_system_trust_store - self.encrypted_credential = encrypted_credential - - -class GoogleBigQueryObjectDataset(Dataset): - """Google BigQuery service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using database + table - properties instead. - :type table_name: any - :param table: The table name of the Google BigQuery. Type: string (or Expression with - resultType string). - :type table: any - :param dataset: The database name of the Google BigQuery. Type: string (or Expression with - resultType string). - :type dataset: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'dataset': {'key': 'typeProperties.dataset', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - table: Optional[Any] = None, - dataset: Optional[Any] = None, - **kwargs - ): - super(GoogleBigQueryObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'GoogleBigQueryObject' # type: str - self.table_name = table_name - self.table = table - self.dataset = dataset - - -class GoogleBigQuerySource(TabularSource): - """A copy activity Google BigQuery service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(GoogleBigQuerySource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'GoogleBigQuerySource' # type: str - self.query = query - - -class GoogleCloudStorageLinkedService(LinkedService): - """Linked service for Google Cloud Storage. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param access_key_id: The access key identifier of the Google Cloud Storage Identity and Access - Management (IAM) user. Type: string (or Expression with resultType string). - :type access_key_id: any - :param secret_access_key: The secret access key of the Google Cloud Storage Identity and Access - Management (IAM) user. - :type secret_access_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param service_url: This value specifies the endpoint to access with the Google Cloud Storage - Connector. This is an optional property; change it only if you want to try a different service - endpoint or want to switch between https and http. Type: string (or Expression with resultType - string). - :type service_url: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'access_key_id': {'key': 'typeProperties.accessKeyId', 'type': 'object'}, - 'secret_access_key': {'key': 'typeProperties.secretAccessKey', 'type': 'SecretBase'}, - 'service_url': {'key': 'typeProperties.serviceUrl', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - access_key_id: Optional[Any] = None, - secret_access_key: Optional["SecretBase"] = None, - service_url: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(GoogleCloudStorageLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'GoogleCloudStorage' # type: str - self.access_key_id = access_key_id - self.secret_access_key = secret_access_key - self.service_url = service_url - self.encrypted_credential = encrypted_credential - - -class GoogleCloudStorageLocation(DatasetLocation): - """The location of Google Cloud Storage dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - :param bucket_name: Specify the bucketName of Google Cloud Storage. Type: string (or Expression - with resultType string). - :type bucket_name: any - :param version: Specify the version of Google Cloud Storage. Type: string (or Expression with - resultType string). - :type version: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - 'bucket_name': {'key': 'bucketName', 'type': 'object'}, - 'version': {'key': 'version', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - folder_path: Optional[Any] = None, - file_name: Optional[Any] = None, - bucket_name: Optional[Any] = None, - version: Optional[Any] = None, - **kwargs - ): - super(GoogleCloudStorageLocation, self).__init__(additional_properties=additional_properties, folder_path=folder_path, file_name=file_name, **kwargs) - self.type = 'GoogleCloudStorageLocation' # type: str - self.bucket_name = bucket_name - self.version = version - - -class GoogleCloudStorageReadSettings(StoreReadSettings): - """Google Cloud Storage read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: Google Cloud Storage wildcardFolderPath. Type: string (or - Expression with resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: Google Cloud Storage wildcardFileName. Type: string (or Expression - with resultType string). - :type wildcard_file_name: any - :param prefix: The prefix filter for the Google Cloud Storage object name. Type: string (or - Expression with resultType string). - :type prefix: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'prefix': {'key': 'prefix', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - 'modified_datetime_start': {'key': 'modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'modifiedDatetimeEnd', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - recursive: Optional[Any] = None, - wildcard_folder_path: Optional[Any] = None, - wildcard_file_name: Optional[Any] = None, - prefix: Optional[Any] = None, - file_list_path: Optional[Any] = None, - enable_partition_discovery: Optional[bool] = None, - partition_root_path: Optional[Any] = None, - delete_files_after_completion: Optional[Any] = None, - modified_datetime_start: Optional[Any] = None, - modified_datetime_end: Optional[Any] = None, - **kwargs - ): - super(GoogleCloudStorageReadSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'GoogleCloudStorageReadSettings' # type: str - self.recursive = recursive - self.wildcard_folder_path = wildcard_folder_path - self.wildcard_file_name = wildcard_file_name - self.prefix = prefix - self.file_list_path = file_list_path - self.enable_partition_discovery = enable_partition_discovery - self.partition_root_path = partition_root_path - self.delete_files_after_completion = delete_files_after_completion - self.modified_datetime_start = modified_datetime_start - self.modified_datetime_end = modified_datetime_end - - -class GreenplumLinkedService(LinkedService): - """Greenplum Database linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: An ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_string: Optional[Any] = None, - pwd: Optional["AzureKeyVaultSecretReference"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(GreenplumLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Greenplum' # type: str - self.connection_string = connection_string - self.pwd = pwd - self.encrypted_credential = encrypted_credential - - -class GreenplumSource(TabularSource): - """A copy activity Greenplum Database source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(GreenplumSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'GreenplumSource' # type: str - self.query = query - - -class GreenplumTableDataset(Dataset): - """Greenplum Database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The table name of Greenplum. Type: string (or Expression with resultType string). - :type table: any - :param schema_type_properties_schema: The schema name of Greenplum. Type: string (or Expression - with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - table: Optional[Any] = None, - schema_type_properties_schema: Optional[Any] = None, - **kwargs - ): - super(GreenplumTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'GreenplumTable' # type: str - self.table_name = table_name - self.table = table - self.schema_type_properties_schema = schema_type_properties_schema - - -class HBaseLinkedService(LinkedService): - """HBase server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The IP address or host name of the HBase server. (i.e. 192.168.222.160). - :type host: any - :param port: The TCP port that the HBase instance uses to listen for client connections. The - default value is 9090. - :type port: any - :param http_path: The partial URL corresponding to the HBase server. (i.e. - /gateway/sandbox/hbase/version). - :type http_path: any - :param authentication_type: Required. The authentication mechanism to use to connect to the - HBase server. Possible values include: "Anonymous", "Basic". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.HBaseAuthenticationType - :param username: The user name used to connect to the HBase instance. - :type username: any - :param password: The password corresponding to the user name. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The - default value is false. - :type enable_ssl: any - :param trusted_cert_path: The full path of the .pem file containing trusted CA certificates for - verifying the server when connecting over SSL. This property can only be set when using SSL on - self-hosted IR. The default value is the cacerts.pem file installed with the IR. - :type trusted_cert_path: any - :param allow_host_name_cn_mismatch: Specifies whether to require a CA-issued SSL certificate - name to match the host name of the server when connecting over SSL. The default value is false. - :type allow_host_name_cn_mismatch: any - :param allow_self_signed_server_cert: Specifies whether to allow self-signed certificates from - the server. The default value is false. - :type allow_self_signed_server_cert: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'http_path': {'key': 'typeProperties.httpPath', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, - 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, - 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, - 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - host: Any, - authentication_type: Union[str, "HBaseAuthenticationType"], - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - port: Optional[Any] = None, - http_path: Optional[Any] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - enable_ssl: Optional[Any] = None, - trusted_cert_path: Optional[Any] = None, - allow_host_name_cn_mismatch: Optional[Any] = None, - allow_self_signed_server_cert: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(HBaseLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'HBase' # type: str - self.host = host - self.port = port - self.http_path = http_path - self.authentication_type = authentication_type - self.username = username - self.password = password - self.enable_ssl = enable_ssl - self.trusted_cert_path = trusted_cert_path - self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch - self.allow_self_signed_server_cert = allow_self_signed_server_cert - self.encrypted_credential = encrypted_credential - - -class HBaseObjectDataset(Dataset): - """HBase server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(HBaseObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'HBaseObject' # type: str - self.table_name = table_name - - -class HBaseSource(TabularSource): - """A copy activity HBase server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(HBaseSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'HBaseSource' # type: str - self.query = query - - -class HdfsLinkedService(LinkedService): - """Hadoop Distributed File System (HDFS) linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param url: Required. The URL of the HDFS service endpoint, e.g. - http://myhostname:50070/webhdfs/v1 . Type: string (or Expression with resultType string). - :type url: any - :param authentication_type: Type of authentication used to connect to the HDFS. Possible values - are: Anonymous and Windows. Type: string (or Expression with resultType string). - :type authentication_type: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - :param user_name: User name for Windows authentication. Type: string (or Expression with - resultType string). - :type user_name: any - :param password: Password for Windows authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - } - - def __init__( - self, - *, - url: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - authentication_type: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - user_name: Optional[Any] = None, - password: Optional["SecretBase"] = None, - **kwargs - ): - super(HdfsLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Hdfs' # type: str - self.url = url - self.authentication_type = authentication_type - self.encrypted_credential = encrypted_credential - self.user_name = user_name - self.password = password - - -class HdfsLocation(DatasetLocation): - """The location of HDFS. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - folder_path: Optional[Any] = None, - file_name: Optional[Any] = None, - **kwargs - ): - super(HdfsLocation, self).__init__(additional_properties=additional_properties, folder_path=folder_path, file_name=file_name, **kwargs) - self.type = 'HdfsLocation' # type: str - - -class HdfsReadSettings(StoreReadSettings): - """HDFS read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: HDFS wildcardFolderPath. Type: string (or Expression with - resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: HDFS wildcardFileName. Type: string (or Expression with resultType - string). - :type wildcard_file_name: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - :param distcp_settings: Specifies Distcp-related settings. - :type distcp_settings: ~azure.synapse.artifacts.v2020_12_01.models.DistcpSettings - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'modified_datetime_start': {'key': 'modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'modifiedDatetimeEnd', 'type': 'object'}, - 'distcp_settings': {'key': 'distcpSettings', 'type': 'DistcpSettings'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - recursive: Optional[Any] = None, - wildcard_folder_path: Optional[Any] = None, - wildcard_file_name: Optional[Any] = None, - file_list_path: Optional[Any] = None, - enable_partition_discovery: Optional[bool] = None, - partition_root_path: Optional[Any] = None, - modified_datetime_start: Optional[Any] = None, - modified_datetime_end: Optional[Any] = None, - distcp_settings: Optional["DistcpSettings"] = None, - delete_files_after_completion: Optional[Any] = None, - **kwargs - ): - super(HdfsReadSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'HdfsReadSettings' # type: str - self.recursive = recursive - self.wildcard_folder_path = wildcard_folder_path - self.wildcard_file_name = wildcard_file_name - self.file_list_path = file_list_path - self.enable_partition_discovery = enable_partition_discovery - self.partition_root_path = partition_root_path - self.modified_datetime_start = modified_datetime_start - self.modified_datetime_end = modified_datetime_end - self.distcp_settings = distcp_settings - self.delete_files_after_completion = delete_files_after_completion - - -class HdfsSource(CopySource): - """A copy activity HDFS source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param distcp_settings: Specifies Distcp-related settings. - :type distcp_settings: ~azure.synapse.artifacts.v2020_12_01.models.DistcpSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'distcp_settings': {'key': 'distcpSettings', 'type': 'DistcpSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - recursive: Optional[Any] = None, - distcp_settings: Optional["DistcpSettings"] = None, - **kwargs - ): - super(HdfsSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'HdfsSource' # type: str - self.recursive = recursive - self.distcp_settings = distcp_settings - - -class HDInsightHiveActivity(ExecutionActivity): - """HDInsight Hive activity type. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param storage_linked_services: Storage linked service references. - :type storage_linked_services: - list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference] - :param arguments: User specified arguments to HDInsightActivity. - :type arguments: list[any] - :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2020_12_01.models.HDInsightActivityDebugInfoOption - :param script_path: Script path. Type: string (or Expression with resultType string). - :type script_path: any - :param script_linked_service: Script linked service reference. - :type script_linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param defines: Allows user to specify defines for Hive job request. - :type defines: dict[str, any] - :param variables: User specified arguments under hivevar namespace. - :type variables: list[any] - :param query_timeout: Query timeout value (in minutes). Effective when the HDInsight cluster - is with ESP (Enterprise Security Package). - :type query_timeout: int - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'storage_linked_services': {'key': 'typeProperties.storageLinkedServices', 'type': '[LinkedServiceReference]'}, - 'arguments': {'key': 'typeProperties.arguments', 'type': '[object]'}, - 'get_debug_info': {'key': 'typeProperties.getDebugInfo', 'type': 'str'}, - 'script_path': {'key': 'typeProperties.scriptPath', 'type': 'object'}, - 'script_linked_service': {'key': 'typeProperties.scriptLinkedService', 'type': 'LinkedServiceReference'}, - 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, - 'variables': {'key': 'typeProperties.variables', 'type': '[object]'}, - 'query_timeout': {'key': 'typeProperties.queryTimeout', 'type': 'int'}, - } - - def __init__( - self, - *, - name: str, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - storage_linked_services: Optional[List["LinkedServiceReference"]] = None, - arguments: Optional[List[Any]] = None, - get_debug_info: Optional[Union[str, "HDInsightActivityDebugInfoOption"]] = None, - script_path: Optional[Any] = None, - script_linked_service: Optional["LinkedServiceReference"] = None, - defines: Optional[Dict[str, Any]] = None, - variables: Optional[List[Any]] = None, - query_timeout: Optional[int] = None, - **kwargs - ): - super(HDInsightHiveActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'HDInsightHive' # type: str - self.storage_linked_services = storage_linked_services - self.arguments = arguments - self.get_debug_info = get_debug_info - self.script_path = script_path - self.script_linked_service = script_linked_service - self.defines = defines - self.variables = variables - self.query_timeout = query_timeout - - -class HDInsightLinkedService(LinkedService): - """HDInsight linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param cluster_uri: Required. HDInsight cluster URI. Type: string (or Expression with - resultType string). - :type cluster_uri: any - :param user_name: HDInsight cluster user name. Type: string (or Expression with resultType - string). - :type user_name: any - :param password: HDInsight cluster password. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param linked_service_name: The Azure Storage linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param hcatalog_linked_service_name: A reference to the Azure SQL linked service that points to - the HCatalog database. - :type hcatalog_linked_service_name: - ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - :param is_esp_enabled: Specify if the HDInsight is created with ESP (Enterprise Security - Package). Type: Boolean. - :type is_esp_enabled: any - :param file_system: Specify the FileSystem if the main storage for the HDInsight is ADLS Gen2. - Type: string (or Expression with resultType string). - :type file_system: any - """ - - _validation = { - 'type': {'required': True}, - 'cluster_uri': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'cluster_uri': {'key': 'typeProperties.clusterUri', 'type': 'object'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'linked_service_name': {'key': 'typeProperties.linkedServiceName', 'type': 'LinkedServiceReference'}, - 'hcatalog_linked_service_name': {'key': 'typeProperties.hcatalogLinkedServiceName', 'type': 'LinkedServiceReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - 'is_esp_enabled': {'key': 'typeProperties.isEspEnabled', 'type': 'object'}, - 'file_system': {'key': 'typeProperties.fileSystem', 'type': 'object'}, - } - - def __init__( - self, - *, - cluster_uri: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - user_name: Optional[Any] = None, - password: Optional["SecretBase"] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - hcatalog_linked_service_name: Optional["LinkedServiceReference"] = None, - encrypted_credential: Optional[Any] = None, - is_esp_enabled: Optional[Any] = None, - file_system: Optional[Any] = None, - **kwargs - ): - super(HDInsightLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'HDInsight' # type: str - self.cluster_uri = cluster_uri - self.user_name = user_name - self.password = password - self.linked_service_name = linked_service_name - self.hcatalog_linked_service_name = hcatalog_linked_service_name - self.encrypted_credential = encrypted_credential - self.is_esp_enabled = is_esp_enabled - self.file_system = file_system - - -class HDInsightMapReduceActivity(ExecutionActivity): - """HDInsight MapReduce activity type. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param storage_linked_services: Storage linked service references. - :type storage_linked_services: - list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference] - :param arguments: User specified arguments to HDInsightActivity. - :type arguments: list[any] - :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2020_12_01.models.HDInsightActivityDebugInfoOption - :param class_name: Required. Class name. Type: string (or Expression with resultType string). - :type class_name: any - :param jar_file_path: Required. Jar path. Type: string (or Expression with resultType string). - :type jar_file_path: any - :param jar_linked_service: Jar linked service reference. - :type jar_linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param jar_libs: Jar libs. - :type jar_libs: list[any] - :param defines: Allows user to specify defines for the MapReduce job request. - :type defines: dict[str, any] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'class_name': {'required': True}, - 'jar_file_path': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'storage_linked_services': {'key': 'typeProperties.storageLinkedServices', 'type': '[LinkedServiceReference]'}, - 'arguments': {'key': 'typeProperties.arguments', 'type': '[object]'}, - 'get_debug_info': {'key': 'typeProperties.getDebugInfo', 'type': 'str'}, - 'class_name': {'key': 'typeProperties.className', 'type': 'object'}, - 'jar_file_path': {'key': 'typeProperties.jarFilePath', 'type': 'object'}, - 'jar_linked_service': {'key': 'typeProperties.jarLinkedService', 'type': 'LinkedServiceReference'}, - 'jar_libs': {'key': 'typeProperties.jarLibs', 'type': '[object]'}, - 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, - } - - def __init__( - self, - *, - name: str, - class_name: Any, - jar_file_path: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - storage_linked_services: Optional[List["LinkedServiceReference"]] = None, - arguments: Optional[List[Any]] = None, - get_debug_info: Optional[Union[str, "HDInsightActivityDebugInfoOption"]] = None, - jar_linked_service: Optional["LinkedServiceReference"] = None, - jar_libs: Optional[List[Any]] = None, - defines: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(HDInsightMapReduceActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'HDInsightMapReduce' # type: str - self.storage_linked_services = storage_linked_services - self.arguments = arguments - self.get_debug_info = get_debug_info - self.class_name = class_name - self.jar_file_path = jar_file_path - self.jar_linked_service = jar_linked_service - self.jar_libs = jar_libs - self.defines = defines - - -class HDInsightOnDemandLinkedService(LinkedService): - """HDInsight ondemand linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param cluster_size: Required. Number of worker/data nodes in the cluster. Suggestion value: 4. - Type: string (or Expression with resultType string). - :type cluster_size: any - :param time_to_live: Required. The allowed idle time for the on-demand HDInsight cluster. - Specifies how long the on-demand HDInsight cluster stays alive after completion of an activity - run if there are no other active jobs in the cluster. The minimum value is 5 mins. Type: string - (or Expression with resultType string). - :type time_to_live: any - :param version: Required. Version of the HDInsight cluster.  Type: string (or Expression with - resultType string). - :type version: any - :param linked_service_name: Required. Azure Storage linked service to be used by the on-demand - cluster for storing and processing data. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param host_subscription_id: Required. The customer’s subscription to host the cluster. Type: - string (or Expression with resultType string). - :type host_subscription_id: any - :param service_principal_id: The service principal id for the hostSubscriptionId. Type: string - (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: The key for the service principal id. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: Required. The Tenant id/name to which the service principal belongs. Type: - string (or Expression with resultType string). - :type tenant: any - :param cluster_resource_group: Required. The resource group where the cluster belongs. Type: - string (or Expression with resultType string). - :type cluster_resource_group: any - :param cluster_name_prefix: The prefix of cluster name, postfix will be distinct with - timestamp. Type: string (or Expression with resultType string). - :type cluster_name_prefix: any - :param cluster_user_name: The username to access the cluster. Type: string (or Expression with - resultType string). - :type cluster_user_name: any - :param cluster_password: The password to access the cluster. - :type cluster_password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param cluster_ssh_user_name: The username to SSH remotely connect to cluster’s node (for - Linux). Type: string (or Expression with resultType string). - :type cluster_ssh_user_name: any - :param cluster_ssh_password: The password to SSH remotely connect cluster’s node (for Linux). - :type cluster_ssh_password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param additional_linked_service_names: Specifies additional storage accounts for the HDInsight - linked service so that the Data Factory service can register them on your behalf. - :type additional_linked_service_names: - list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference] - :param hcatalog_linked_service_name: The name of Azure SQL linked service that point to the - HCatalog database. The on-demand HDInsight cluster is created by using the Azure SQL database - as the metastore. - :type hcatalog_linked_service_name: - ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param cluster_type: The cluster type. Type: string (or Expression with resultType string). - :type cluster_type: any - :param spark_version: The version of spark if the cluster type is 'spark'. Type: string (or - Expression with resultType string). - :type spark_version: any - :param core_configuration: Specifies the core configuration parameters (as in core-site.xml) - for the HDInsight cluster to be created. - :type core_configuration: any - :param h_base_configuration: Specifies the HBase configuration parameters (hbase-site.xml) for - the HDInsight cluster. - :type h_base_configuration: any - :param hdfs_configuration: Specifies the HDFS configuration parameters (hdfs-site.xml) for the - HDInsight cluster. - :type hdfs_configuration: any - :param hive_configuration: Specifies the hive configuration parameters (hive-site.xml) for the - HDInsight cluster. - :type hive_configuration: any - :param map_reduce_configuration: Specifies the MapReduce configuration parameters - (mapred-site.xml) for the HDInsight cluster. - :type map_reduce_configuration: any - :param oozie_configuration: Specifies the Oozie configuration parameters (oozie-site.xml) for - the HDInsight cluster. - :type oozie_configuration: any - :param storm_configuration: Specifies the Storm configuration parameters (storm-site.xml) for - the HDInsight cluster. - :type storm_configuration: any - :param yarn_configuration: Specifies the Yarn configuration parameters (yarn-site.xml) for the - HDInsight cluster. - :type yarn_configuration: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - :param head_node_size: Specifies the size of the head node for the HDInsight cluster. - :type head_node_size: any - :param data_node_size: Specifies the size of the data node for the HDInsight cluster. - :type data_node_size: any - :param zookeeper_node_size: Specifies the size of the Zoo Keeper node for the HDInsight - cluster. - :type zookeeper_node_size: any - :param script_actions: Custom script actions to run on HDI ondemand cluster once it's up. - Please refer to - https://docs.microsoft.com/en-us/azure/hdinsight/hdinsight-hadoop-customize-cluster-linux?toc=%2Fen-us%2Fazure%2Fhdinsight%2Fr-server%2FTOC.json&bc=%2Fen-us%2Fazure%2Fbread%2Ftoc.json#understanding-script-actions. - :type script_actions: list[~azure.synapse.artifacts.v2020_12_01.models.ScriptAction] - :param virtual_network_id: The ARM resource ID for the vNet to which the cluster should be - joined after creation. Type: string (or Expression with resultType string). - :type virtual_network_id: any - :param subnet_name: The ARM resource ID for the subnet in the vNet. If virtualNetworkId was - specified, then this property is required. Type: string (or Expression with resultType string). - :type subnet_name: any - """ - - _validation = { - 'type': {'required': True}, - 'cluster_size': {'required': True}, - 'time_to_live': {'required': True}, - 'version': {'required': True}, - 'linked_service_name': {'required': True}, - 'host_subscription_id': {'required': True}, - 'tenant': {'required': True}, - 'cluster_resource_group': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'cluster_size': {'key': 'typeProperties.clusterSize', 'type': 'object'}, - 'time_to_live': {'key': 'typeProperties.timeToLive', 'type': 'object'}, - 'version': {'key': 'typeProperties.version', 'type': 'object'}, - 'linked_service_name': {'key': 'typeProperties.linkedServiceName', 'type': 'LinkedServiceReference'}, - 'host_subscription_id': {'key': 'typeProperties.hostSubscriptionId', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'cluster_resource_group': {'key': 'typeProperties.clusterResourceGroup', 'type': 'object'}, - 'cluster_name_prefix': {'key': 'typeProperties.clusterNamePrefix', 'type': 'object'}, - 'cluster_user_name': {'key': 'typeProperties.clusterUserName', 'type': 'object'}, - 'cluster_password': {'key': 'typeProperties.clusterPassword', 'type': 'SecretBase'}, - 'cluster_ssh_user_name': {'key': 'typeProperties.clusterSshUserName', 'type': 'object'}, - 'cluster_ssh_password': {'key': 'typeProperties.clusterSshPassword', 'type': 'SecretBase'}, - 'additional_linked_service_names': {'key': 'typeProperties.additionalLinkedServiceNames', 'type': '[LinkedServiceReference]'}, - 'hcatalog_linked_service_name': {'key': 'typeProperties.hcatalogLinkedServiceName', 'type': 'LinkedServiceReference'}, - 'cluster_type': {'key': 'typeProperties.clusterType', 'type': 'object'}, - 'spark_version': {'key': 'typeProperties.sparkVersion', 'type': 'object'}, - 'core_configuration': {'key': 'typeProperties.coreConfiguration', 'type': 'object'}, - 'h_base_configuration': {'key': 'typeProperties.hBaseConfiguration', 'type': 'object'}, - 'hdfs_configuration': {'key': 'typeProperties.hdfsConfiguration', 'type': 'object'}, - 'hive_configuration': {'key': 'typeProperties.hiveConfiguration', 'type': 'object'}, - 'map_reduce_configuration': {'key': 'typeProperties.mapReduceConfiguration', 'type': 'object'}, - 'oozie_configuration': {'key': 'typeProperties.oozieConfiguration', 'type': 'object'}, - 'storm_configuration': {'key': 'typeProperties.stormConfiguration', 'type': 'object'}, - 'yarn_configuration': {'key': 'typeProperties.yarnConfiguration', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - 'head_node_size': {'key': 'typeProperties.headNodeSize', 'type': 'object'}, - 'data_node_size': {'key': 'typeProperties.dataNodeSize', 'type': 'object'}, - 'zookeeper_node_size': {'key': 'typeProperties.zookeeperNodeSize', 'type': 'object'}, - 'script_actions': {'key': 'typeProperties.scriptActions', 'type': '[ScriptAction]'}, - 'virtual_network_id': {'key': 'typeProperties.virtualNetworkId', 'type': 'object'}, - 'subnet_name': {'key': 'typeProperties.subnetName', 'type': 'object'}, - } - - def __init__( - self, - *, - cluster_size: Any, - time_to_live: Any, - version: Any, - linked_service_name: "LinkedServiceReference", - host_subscription_id: Any, - tenant: Any, - cluster_resource_group: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - service_principal_id: Optional[Any] = None, - service_principal_key: Optional["SecretBase"] = None, - cluster_name_prefix: Optional[Any] = None, - cluster_user_name: Optional[Any] = None, - cluster_password: Optional["SecretBase"] = None, - cluster_ssh_user_name: Optional[Any] = None, - cluster_ssh_password: Optional["SecretBase"] = None, - additional_linked_service_names: Optional[List["LinkedServiceReference"]] = None, - hcatalog_linked_service_name: Optional["LinkedServiceReference"] = None, - cluster_type: Optional[Any] = None, - spark_version: Optional[Any] = None, - core_configuration: Optional[Any] = None, - h_base_configuration: Optional[Any] = None, - hdfs_configuration: Optional[Any] = None, - hive_configuration: Optional[Any] = None, - map_reduce_configuration: Optional[Any] = None, - oozie_configuration: Optional[Any] = None, - storm_configuration: Optional[Any] = None, - yarn_configuration: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - head_node_size: Optional[Any] = None, - data_node_size: Optional[Any] = None, - zookeeper_node_size: Optional[Any] = None, - script_actions: Optional[List["ScriptAction"]] = None, - virtual_network_id: Optional[Any] = None, - subnet_name: Optional[Any] = None, - **kwargs - ): - super(HDInsightOnDemandLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'HDInsightOnDemand' # type: str - self.cluster_size = cluster_size - self.time_to_live = time_to_live - self.version = version - self.linked_service_name = linked_service_name - self.host_subscription_id = host_subscription_id - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.cluster_resource_group = cluster_resource_group - self.cluster_name_prefix = cluster_name_prefix - self.cluster_user_name = cluster_user_name - self.cluster_password = cluster_password - self.cluster_ssh_user_name = cluster_ssh_user_name - self.cluster_ssh_password = cluster_ssh_password - self.additional_linked_service_names = additional_linked_service_names - self.hcatalog_linked_service_name = hcatalog_linked_service_name - self.cluster_type = cluster_type - self.spark_version = spark_version - self.core_configuration = core_configuration - self.h_base_configuration = h_base_configuration - self.hdfs_configuration = hdfs_configuration - self.hive_configuration = hive_configuration - self.map_reduce_configuration = map_reduce_configuration - self.oozie_configuration = oozie_configuration - self.storm_configuration = storm_configuration - self.yarn_configuration = yarn_configuration - self.encrypted_credential = encrypted_credential - self.head_node_size = head_node_size - self.data_node_size = data_node_size - self.zookeeper_node_size = zookeeper_node_size - self.script_actions = script_actions - self.virtual_network_id = virtual_network_id - self.subnet_name = subnet_name - - -class HDInsightPigActivity(ExecutionActivity): - """HDInsight Pig activity type. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param storage_linked_services: Storage linked service references. - :type storage_linked_services: - list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference] - :param arguments: User specified arguments to HDInsightActivity. Type: array (or Expression - with resultType array). - :type arguments: any - :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2020_12_01.models.HDInsightActivityDebugInfoOption - :param script_path: Script path. Type: string (or Expression with resultType string). - :type script_path: any - :param script_linked_service: Script linked service reference. - :type script_linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param defines: Allows user to specify defines for Pig job request. - :type defines: dict[str, any] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'storage_linked_services': {'key': 'typeProperties.storageLinkedServices', 'type': '[LinkedServiceReference]'}, - 'arguments': {'key': 'typeProperties.arguments', 'type': 'object'}, - 'get_debug_info': {'key': 'typeProperties.getDebugInfo', 'type': 'str'}, - 'script_path': {'key': 'typeProperties.scriptPath', 'type': 'object'}, - 'script_linked_service': {'key': 'typeProperties.scriptLinkedService', 'type': 'LinkedServiceReference'}, - 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, - } - - def __init__( - self, - *, - name: str, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - storage_linked_services: Optional[List["LinkedServiceReference"]] = None, - arguments: Optional[Any] = None, - get_debug_info: Optional[Union[str, "HDInsightActivityDebugInfoOption"]] = None, - script_path: Optional[Any] = None, - script_linked_service: Optional["LinkedServiceReference"] = None, - defines: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(HDInsightPigActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'HDInsightPig' # type: str - self.storage_linked_services = storage_linked_services - self.arguments = arguments - self.get_debug_info = get_debug_info - self.script_path = script_path - self.script_linked_service = script_linked_service - self.defines = defines - - -class HDInsightSparkActivity(ExecutionActivity): - """HDInsight Spark activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param root_path: Required. The root path in 'sparkJobLinkedService' for all the job’s files. - Type: string (or Expression with resultType string). - :type root_path: any - :param entry_file_path: Required. The relative path to the root folder of the code/package to - be executed. Type: string (or Expression with resultType string). - :type entry_file_path: any - :param arguments: The user-specified arguments to HDInsightSparkActivity. - :type arguments: list[any] - :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2020_12_01.models.HDInsightActivityDebugInfoOption - :param spark_job_linked_service: The storage linked service for uploading the entry file and - dependencies, and for receiving logs. - :type spark_job_linked_service: - ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param class_name: The application's Java/Spark main class. - :type class_name: str - :param proxy_user: The user to impersonate that will execute the job. Type: string (or - Expression with resultType string). - :type proxy_user: any - :param spark_config: Spark configuration property. - :type spark_config: dict[str, any] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'root_path': {'required': True}, - 'entry_file_path': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'root_path': {'key': 'typeProperties.rootPath', 'type': 'object'}, - 'entry_file_path': {'key': 'typeProperties.entryFilePath', 'type': 'object'}, - 'arguments': {'key': 'typeProperties.arguments', 'type': '[object]'}, - 'get_debug_info': {'key': 'typeProperties.getDebugInfo', 'type': 'str'}, - 'spark_job_linked_service': {'key': 'typeProperties.sparkJobLinkedService', 'type': 'LinkedServiceReference'}, - 'class_name': {'key': 'typeProperties.className', 'type': 'str'}, - 'proxy_user': {'key': 'typeProperties.proxyUser', 'type': 'object'}, - 'spark_config': {'key': 'typeProperties.sparkConfig', 'type': '{object}'}, - } - - def __init__( - self, - *, - name: str, - root_path: Any, - entry_file_path: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - arguments: Optional[List[Any]] = None, - get_debug_info: Optional[Union[str, "HDInsightActivityDebugInfoOption"]] = None, - spark_job_linked_service: Optional["LinkedServiceReference"] = None, - class_name: Optional[str] = None, - proxy_user: Optional[Any] = None, - spark_config: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(HDInsightSparkActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'HDInsightSpark' # type: str - self.root_path = root_path - self.entry_file_path = entry_file_path - self.arguments = arguments - self.get_debug_info = get_debug_info - self.spark_job_linked_service = spark_job_linked_service - self.class_name = class_name - self.proxy_user = proxy_user - self.spark_config = spark_config - - -class HDInsightStreamingActivity(ExecutionActivity): - """HDInsight streaming activity type. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param storage_linked_services: Storage linked service references. - :type storage_linked_services: - list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference] - :param arguments: User specified arguments to HDInsightActivity. - :type arguments: list[any] - :param get_debug_info: Debug info option. Possible values include: "None", "Always", "Failure". - :type get_debug_info: str or - ~azure.synapse.artifacts.v2020_12_01.models.HDInsightActivityDebugInfoOption - :param mapper: Required. Mapper executable name. Type: string (or Expression with resultType - string). - :type mapper: any - :param reducer: Required. Reducer executable name. Type: string (or Expression with resultType - string). - :type reducer: any - :param input: Required. Input blob path. Type: string (or Expression with resultType string). - :type input: any - :param output: Required. Output blob path. Type: string (or Expression with resultType string). - :type output: any - :param file_paths: Required. Paths to streaming job files. Can be directories. - :type file_paths: list[any] - :param file_linked_service: Linked service reference where the files are located. - :type file_linked_service: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param combiner: Combiner executable name. Type: string (or Expression with resultType string). - :type combiner: any - :param command_environment: Command line environment values. - :type command_environment: list[any] - :param defines: Allows user to specify defines for streaming job request. - :type defines: dict[str, any] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'mapper': {'required': True}, - 'reducer': {'required': True}, - 'input': {'required': True}, - 'output': {'required': True}, - 'file_paths': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'storage_linked_services': {'key': 'typeProperties.storageLinkedServices', 'type': '[LinkedServiceReference]'}, - 'arguments': {'key': 'typeProperties.arguments', 'type': '[object]'}, - 'get_debug_info': {'key': 'typeProperties.getDebugInfo', 'type': 'str'}, - 'mapper': {'key': 'typeProperties.mapper', 'type': 'object'}, - 'reducer': {'key': 'typeProperties.reducer', 'type': 'object'}, - 'input': {'key': 'typeProperties.input', 'type': 'object'}, - 'output': {'key': 'typeProperties.output', 'type': 'object'}, - 'file_paths': {'key': 'typeProperties.filePaths', 'type': '[object]'}, - 'file_linked_service': {'key': 'typeProperties.fileLinkedService', 'type': 'LinkedServiceReference'}, - 'combiner': {'key': 'typeProperties.combiner', 'type': 'object'}, - 'command_environment': {'key': 'typeProperties.commandEnvironment', 'type': '[object]'}, - 'defines': {'key': 'typeProperties.defines', 'type': '{object}'}, - } - - def __init__( - self, - *, - name: str, - mapper: Any, - reducer: Any, - input: Any, - output: Any, - file_paths: List[Any], - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - storage_linked_services: Optional[List["LinkedServiceReference"]] = None, - arguments: Optional[List[Any]] = None, - get_debug_info: Optional[Union[str, "HDInsightActivityDebugInfoOption"]] = None, - file_linked_service: Optional["LinkedServiceReference"] = None, - combiner: Optional[Any] = None, - command_environment: Optional[List[Any]] = None, - defines: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(HDInsightStreamingActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'HDInsightStreaming' # type: str - self.storage_linked_services = storage_linked_services - self.arguments = arguments - self.get_debug_info = get_debug_info - self.mapper = mapper - self.reducer = reducer - self.input = input - self.output = output - self.file_paths = file_paths - self.file_linked_service = file_linked_service - self.combiner = combiner - self.command_environment = command_environment - self.defines = defines - - -class HiveLinkedService(LinkedService): - """Hive Server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. IP address or host name of the Hive server, separated by ';' for - multiple hosts (only when serviceDiscoveryMode is enable). - :type host: any - :param port: The TCP port that the Hive server uses to listen for client connections. - :type port: any - :param server_type: The type of Hive server. Possible values include: "HiveServer1", - "HiveServer2", "HiveThriftServer". - :type server_type: str or ~azure.synapse.artifacts.v2020_12_01.models.HiveServerType - :param thrift_transport_protocol: The transport protocol to use in the Thrift layer. Possible - values include: "Binary", "SASL", "HTTP ". - :type thrift_transport_protocol: str or - ~azure.synapse.artifacts.v2020_12_01.models.HiveThriftTransportProtocol - :param authentication_type: Required. The authentication method used to access the Hive server. - Possible values include: "Anonymous", "Username", "UsernameAndPassword", - "WindowsAzureHDInsightService". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.HiveAuthenticationType - :param service_discovery_mode: true to indicate using the ZooKeeper service, false not. - :type service_discovery_mode: any - :param zoo_keeper_name_space: The namespace on ZooKeeper under which Hive Server 2 nodes are - added. - :type zoo_keeper_name_space: any - :param use_native_query: Specifies whether the driver uses native HiveQL queries,or converts - them into an equivalent form in HiveQL. - :type use_native_query: any - :param username: The user name that you use to access Hive Server. - :type username: any - :param password: The password corresponding to the user name that you provided in the Username - field. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param http_path: The partial URL corresponding to the Hive server. - :type http_path: any - :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The - default value is false. - :type enable_ssl: any - :param trusted_cert_path: The full path of the .pem file containing trusted CA certificates for - verifying the server when connecting over SSL. This property can only be set when using SSL on - self-hosted IR. The default value is the cacerts.pem file installed with the IR. - :type trusted_cert_path: any - :param use_system_trust_store: Specifies whether to use a CA certificate from the system trust - store or from a specified PEM file. The default value is false. - :type use_system_trust_store: any - :param allow_host_name_cn_mismatch: Specifies whether to require a CA-issued SSL certificate - name to match the host name of the server when connecting over SSL. The default value is false. - :type allow_host_name_cn_mismatch: any - :param allow_self_signed_server_cert: Specifies whether to allow self-signed certificates from - the server. The default value is false. - :type allow_self_signed_server_cert: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'server_type': {'key': 'typeProperties.serverType', 'type': 'str'}, - 'thrift_transport_protocol': {'key': 'typeProperties.thriftTransportProtocol', 'type': 'str'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'service_discovery_mode': {'key': 'typeProperties.serviceDiscoveryMode', 'type': 'object'}, - 'zoo_keeper_name_space': {'key': 'typeProperties.zooKeeperNameSpace', 'type': 'object'}, - 'use_native_query': {'key': 'typeProperties.useNativeQuery', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'http_path': {'key': 'typeProperties.httpPath', 'type': 'object'}, - 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, - 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, - 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, - 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, - 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - host: Any, - authentication_type: Union[str, "HiveAuthenticationType"], - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - port: Optional[Any] = None, - server_type: Optional[Union[str, "HiveServerType"]] = None, - thrift_transport_protocol: Optional[Union[str, "HiveThriftTransportProtocol"]] = None, - service_discovery_mode: Optional[Any] = None, - zoo_keeper_name_space: Optional[Any] = None, - use_native_query: Optional[Any] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - http_path: Optional[Any] = None, - enable_ssl: Optional[Any] = None, - trusted_cert_path: Optional[Any] = None, - use_system_trust_store: Optional[Any] = None, - allow_host_name_cn_mismatch: Optional[Any] = None, - allow_self_signed_server_cert: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(HiveLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Hive' # type: str - self.host = host - self.port = port - self.server_type = server_type - self.thrift_transport_protocol = thrift_transport_protocol - self.authentication_type = authentication_type - self.service_discovery_mode = service_discovery_mode - self.zoo_keeper_name_space = zoo_keeper_name_space - self.use_native_query = use_native_query - self.username = username - self.password = password - self.http_path = http_path - self.enable_ssl = enable_ssl - self.trusted_cert_path = trusted_cert_path - self.use_system_trust_store = use_system_trust_store - self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch - self.allow_self_signed_server_cert = allow_self_signed_server_cert - self.encrypted_credential = encrypted_credential - - -class HiveObjectDataset(Dataset): - """Hive Server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The table name of the Hive. Type: string (or Expression with resultType string). - :type table: any - :param schema_type_properties_schema: The schema name of the Hive. Type: string (or Expression - with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - table: Optional[Any] = None, - schema_type_properties_schema: Optional[Any] = None, - **kwargs - ): - super(HiveObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'HiveObject' # type: str - self.table_name = table_name - self.table = table - self.schema_type_properties_schema = schema_type_properties_schema - - -class HiveSource(TabularSource): - """A copy activity Hive Server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(HiveSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'HiveSource' # type: str - self.query = query - - -class HttpDataset(Dataset): - """A file in an HTTP web server. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param relative_url: The relative URL based on the URL in the HttpLinkedService refers to an - HTTP file Type: string (or Expression with resultType string). - :type relative_url: any - :param request_method: The HTTP method for the HTTP request. Type: string (or Expression with - resultType string). - :type request_method: any - :param request_body: The body for the HTTP request. Type: string (or Expression with resultType - string). - :type request_body: any - :param additional_headers: The headers for the HTTP Request. e.g. - request-header-name-1:request-header-value-1 - ... - request-header-name-n:request-header-value-n Type: string (or Expression with resultType - string). - :type additional_headers: any - :param format: The format of files. - :type format: ~azure.synapse.artifacts.v2020_12_01.models.DatasetStorageFormat - :param compression: The data compression method used on files. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'relative_url': {'key': 'typeProperties.relativeUrl', 'type': 'object'}, - 'request_method': {'key': 'typeProperties.requestMethod', 'type': 'object'}, - 'request_body': {'key': 'typeProperties.requestBody', 'type': 'object'}, - 'additional_headers': {'key': 'typeProperties.additionalHeaders', 'type': 'object'}, - 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - relative_url: Optional[Any] = None, - request_method: Optional[Any] = None, - request_body: Optional[Any] = None, - additional_headers: Optional[Any] = None, - format: Optional["DatasetStorageFormat"] = None, - compression: Optional["DatasetCompression"] = None, - **kwargs - ): - super(HttpDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'HttpFile' # type: str - self.relative_url = relative_url - self.request_method = request_method - self.request_body = request_body - self.additional_headers = additional_headers - self.format = format - self.compression = compression - - -class HttpLinkedService(LinkedService): - """Linked service for an HTTP source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param url: Required. The base URL of the HTTP endpoint, e.g. http://www.microsoft.com. Type: - string (or Expression with resultType string). - :type url: any - :param authentication_type: The authentication type to be used to connect to the HTTP server. - Possible values include: "Basic", "Anonymous", "Digest", "Windows", "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.HttpAuthenticationType - :param user_name: User name for Basic, Digest, or Windows authentication. Type: string (or - Expression with resultType string). - :type user_name: any - :param password: Password for Basic, Digest, Windows, or ClientCertificate with - EmbeddedCertData authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param embedded_cert_data: Base64 encoded certificate data for ClientCertificate - authentication. For on-premises copy with ClientCertificate authentication, either - CertThumbprint or EmbeddedCertData/Password should be specified. Type: string (or Expression - with resultType string). - :type embedded_cert_data: any - :param cert_thumbprint: Thumbprint of certificate for ClientCertificate authentication. Only - valid for on-premises copy. For on-premises copy with ClientCertificate authentication, either - CertThumbprint or EmbeddedCertData/Password should be specified. Type: string (or Expression - with resultType string). - :type cert_thumbprint: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - :param enable_server_certificate_validation: If true, validate the HTTPS server SSL - certificate. Default value is true. Type: boolean (or Expression with resultType boolean). - :type enable_server_certificate_validation: any - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'embedded_cert_data': {'key': 'typeProperties.embeddedCertData', 'type': 'object'}, - 'cert_thumbprint': {'key': 'typeProperties.certThumbprint', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - 'enable_server_certificate_validation': {'key': 'typeProperties.enableServerCertificateValidation', 'type': 'object'}, - } - - def __init__( - self, - *, - url: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - authentication_type: Optional[Union[str, "HttpAuthenticationType"]] = None, - user_name: Optional[Any] = None, - password: Optional["SecretBase"] = None, - embedded_cert_data: Optional[Any] = None, - cert_thumbprint: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - enable_server_certificate_validation: Optional[Any] = None, - **kwargs - ): - super(HttpLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'HttpServer' # type: str - self.url = url - self.authentication_type = authentication_type - self.user_name = user_name - self.password = password - self.embedded_cert_data = embedded_cert_data - self.cert_thumbprint = cert_thumbprint - self.encrypted_credential = encrypted_credential - self.enable_server_certificate_validation = enable_server_certificate_validation - - -class HttpReadSettings(StoreReadSettings): - """Sftp read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param request_method: The HTTP method used to call the RESTful API. The default is GET. Type: - string (or Expression with resultType string). - :type request_method: any - :param request_body: The HTTP request body to the RESTful API if requestMethod is POST. Type: - string (or Expression with resultType string). - :type request_body: any - :param additional_headers: The additional HTTP headers in the request to the RESTful API. Type: - string (or Expression with resultType string). - :type additional_headers: any - :param request_timeout: Specifies the timeout for a HTTP client to get HTTP response from HTTP - server. - :type request_timeout: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'request_method': {'key': 'requestMethod', 'type': 'object'}, - 'request_body': {'key': 'requestBody', 'type': 'object'}, - 'additional_headers': {'key': 'additionalHeaders', 'type': 'object'}, - 'request_timeout': {'key': 'requestTimeout', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - request_method: Optional[Any] = None, - request_body: Optional[Any] = None, - additional_headers: Optional[Any] = None, - request_timeout: Optional[Any] = None, - enable_partition_discovery: Optional[bool] = None, - partition_root_path: Optional[Any] = None, - **kwargs - ): - super(HttpReadSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'HttpReadSettings' # type: str - self.request_method = request_method - self.request_body = request_body - self.additional_headers = additional_headers - self.request_timeout = request_timeout - self.enable_partition_discovery = enable_partition_discovery - self.partition_root_path = partition_root_path - - -class HttpServerLocation(DatasetLocation): - """The location of http server. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - :param relative_url: Specify the relativeUrl of http server. Type: string (or Expression with - resultType string). - :type relative_url: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - 'relative_url': {'key': 'relativeUrl', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - folder_path: Optional[Any] = None, - file_name: Optional[Any] = None, - relative_url: Optional[Any] = None, - **kwargs - ): - super(HttpServerLocation, self).__init__(additional_properties=additional_properties, folder_path=folder_path, file_name=file_name, **kwargs) - self.type = 'HttpServerLocation' # type: str - self.relative_url = relative_url - - -class HttpSource(CopySource): - """A copy activity source for an HTTP file. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param http_request_timeout: Specifies the timeout for a HTTP client to get HTTP response from - HTTP server. The default value is equivalent to System.Net.HttpWebRequest.Timeout. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - http_request_timeout: Optional[Any] = None, - **kwargs - ): - super(HttpSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'HttpSource' # type: str - self.http_request_timeout = http_request_timeout - - -class HubspotLinkedService(LinkedService): - """Hubspot Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param client_id: Required. The client ID associated with your Hubspot application. - :type client_id: any - :param client_secret: The client secret associated with your Hubspot application. - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param access_token: The access token obtained when initially authenticating your OAuth - integration. - :type access_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param refresh_token: The refresh token obtained when initially authenticating your OAuth - integration. - :type refresh_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'client_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, - 'refresh_token': {'key': 'typeProperties.refreshToken', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - client_id: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - client_secret: Optional["SecretBase"] = None, - access_token: Optional["SecretBase"] = None, - refresh_token: Optional["SecretBase"] = None, - use_encrypted_endpoints: Optional[Any] = None, - use_host_verification: Optional[Any] = None, - use_peer_verification: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(HubspotLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Hubspot' # type: str - self.client_id = client_id - self.client_secret = client_secret - self.access_token = access_token - self.refresh_token = refresh_token - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential - - -class HubspotObjectDataset(Dataset): - """Hubspot Service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(HubspotObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'HubspotObject' # type: str - self.table_name = table_name - - -class HubspotSource(TabularSource): - """A copy activity Hubspot Service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(HubspotSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'HubspotSource' # type: str - self.query = query - - -class IfConditionActivity(ControlActivity): - """This activity evaluates a boolean expression and executes either the activities under the ifTrueActivities property or the ifFalseActivities property depending on the result of the expression. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param expression: Required. An expression that would evaluate to Boolean. This is used to - determine the block of activities (ifTrueActivities or ifFalseActivities) that will be - executed. - :type expression: ~azure.synapse.artifacts.v2020_12_01.models.Expression - :param if_true_activities: List of activities to execute if expression is evaluated to true. - This is an optional property and if not provided, the activity will exit without any action. - :type if_true_activities: list[~azure.synapse.artifacts.v2020_12_01.models.Activity] - :param if_false_activities: List of activities to execute if expression is evaluated to false. - This is an optional property and if not provided, the activity will exit without any action. - :type if_false_activities: list[~azure.synapse.artifacts.v2020_12_01.models.Activity] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'expression': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'expression': {'key': 'typeProperties.expression', 'type': 'Expression'}, - 'if_true_activities': {'key': 'typeProperties.ifTrueActivities', 'type': '[Activity]'}, - 'if_false_activities': {'key': 'typeProperties.ifFalseActivities', 'type': '[Activity]'}, - } - - def __init__( - self, - *, - name: str, - expression: "Expression", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - if_true_activities: Optional[List["Activity"]] = None, - if_false_activities: Optional[List["Activity"]] = None, - **kwargs - ): - super(IfConditionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) - self.type = 'IfCondition' # type: str - self.expression = expression - self.if_true_activities = if_true_activities - self.if_false_activities = if_false_activities - - -class ImpalaLinkedService(LinkedService): - """Impala server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The IP address or host name of the Impala server. (i.e. - 192.168.222.160). - :type host: any - :param port: The TCP port that the Impala server uses to listen for client connections. The - default value is 21050. - :type port: any - :param authentication_type: Required. The authentication type to use. Possible values include: - "Anonymous", "SASLUsername", "UsernameAndPassword". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.ImpalaAuthenticationType - :param username: The user name used to access the Impala server. The default value is anonymous - when using SASLUsername. - :type username: any - :param password: The password corresponding to the user name when using UsernameAndPassword. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The - default value is false. - :type enable_ssl: any - :param trusted_cert_path: The full path of the .pem file containing trusted CA certificates for - verifying the server when connecting over SSL. This property can only be set when using SSL on - self-hosted IR. The default value is the cacerts.pem file installed with the IR. - :type trusted_cert_path: any - :param use_system_trust_store: Specifies whether to use a CA certificate from the system trust - store or from a specified PEM file. The default value is false. - :type use_system_trust_store: any - :param allow_host_name_cn_mismatch: Specifies whether to require a CA-issued SSL certificate - name to match the host name of the server when connecting over SSL. The default value is false. - :type allow_host_name_cn_mismatch: any - :param allow_self_signed_server_cert: Specifies whether to allow self-signed certificates from - the server. The default value is false. - :type allow_self_signed_server_cert: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, - 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, - 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, - 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, - 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - host: Any, - authentication_type: Union[str, "ImpalaAuthenticationType"], - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - port: Optional[Any] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - enable_ssl: Optional[Any] = None, - trusted_cert_path: Optional[Any] = None, - use_system_trust_store: Optional[Any] = None, - allow_host_name_cn_mismatch: Optional[Any] = None, - allow_self_signed_server_cert: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(ImpalaLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Impala' # type: str - self.host = host - self.port = port - self.authentication_type = authentication_type - self.username = username - self.password = password - self.enable_ssl = enable_ssl - self.trusted_cert_path = trusted_cert_path - self.use_system_trust_store = use_system_trust_store - self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch - self.allow_self_signed_server_cert = allow_self_signed_server_cert - self.encrypted_credential = encrypted_credential - - -class ImpalaObjectDataset(Dataset): - """Impala server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The table name of the Impala. Type: string (or Expression with resultType - string). - :type table: any - :param schema_type_properties_schema: The schema name of the Impala. Type: string (or - Expression with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - table: Optional[Any] = None, - schema_type_properties_schema: Optional[Any] = None, - **kwargs - ): - super(ImpalaObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'ImpalaObject' # type: str - self.table_name = table_name - self.table = table - self.schema_type_properties_schema = schema_type_properties_schema - - -class ImpalaSource(TabularSource): - """A copy activity Impala server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(ImpalaSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'ImpalaSource' # type: str - self.query = query - - -class InformixLinkedService(LinkedService): - """Informix linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The non-access credential portion of the connection string - as well as an optional encrypted credential. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param authentication_type: Type of authentication used to connect to the Informix as ODBC data - store. Possible values are: Anonymous and Basic. Type: string (or Expression with resultType - string). - :type authentication_type: any - :param credential: The access credential portion of the connection string specified in - driver-specific property-value format. - :type credential: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param user_name: User name for Basic authentication. Type: string (or Expression with - resultType string). - :type user_name: any - :param password: Password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, - 'credential': {'key': 'typeProperties.credential', 'type': 'SecretBase'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - connection_string: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - authentication_type: Optional[Any] = None, - credential: Optional["SecretBase"] = None, - user_name: Optional[Any] = None, - password: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(InformixLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Informix' # type: str - self.connection_string = connection_string - self.authentication_type = authentication_type - self.credential = credential - self.user_name = user_name - self.password = password - self.encrypted_credential = encrypted_credential - - -class InformixSink(CopySink): - """A copy activity Informix sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param pre_copy_script: A query to execute before starting the copy. Type: string (or - Expression with resultType string). - :type pre_copy_script: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - pre_copy_script: Optional[Any] = None, - **kwargs - ): - super(InformixSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'InformixSink' # type: str - self.pre_copy_script = pre_copy_script - - -class InformixSource(TabularSource): - """A copy activity source for Informix. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(InformixSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'InformixSource' # type: str - self.query = query - - -class InformixTableDataset(Dataset): - """The Informix table dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The Informix table name. Type: string (or Expression with resultType - string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(InformixTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'InformixTable' # type: str - self.table_name = table_name - - -class IntegrationRuntime(msrest.serialization.Model): - """Azure Synapse nested object which serves as a compute resource for activities. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: ManagedIntegrationRuntime, SelfHostedIntegrationRuntime. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of integration runtime.Constant filled by server. Possible values - include: "Managed", "SelfHosted". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeType - :param description: Integration runtime description. - :type description: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - } - - _subtype_map = { - 'type': {'Managed': 'ManagedIntegrationRuntime', 'SelfHosted': 'SelfHostedIntegrationRuntime'} - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - **kwargs - ): - super(IntegrationRuntime, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = 'IntegrationRuntime' # type: str - self.description = description - - -class IntegrationRuntimeComputeProperties(msrest.serialization.Model): - """The compute resource properties for managed integration runtime. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param location: The location for managed integration runtime. The supported regions could be - found on - https://docs.microsoft.com/en-us/azure/data-factory/data-factory-data-movement-activities. - :type location: str - :param node_size: The node size requirement to managed integration runtime. - :type node_size: str - :param number_of_nodes: The required number of nodes for managed integration runtime. - :type number_of_nodes: int - :param max_parallel_executions_per_node: Maximum parallel executions count per node for managed - integration runtime. - :type max_parallel_executions_per_node: int - :param data_flow_properties: Data flow properties for managed integration runtime. - :type data_flow_properties: - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeDataFlowProperties - :param v_net_properties: VNet properties for managed integration runtime. - :type v_net_properties: - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeVNetProperties - """ - - _validation = { - 'number_of_nodes': {'minimum': 1}, - 'max_parallel_executions_per_node': {'minimum': 1}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'location': {'key': 'location', 'type': 'str'}, - 'node_size': {'key': 'nodeSize', 'type': 'str'}, - 'number_of_nodes': {'key': 'numberOfNodes', 'type': 'int'}, - 'max_parallel_executions_per_node': {'key': 'maxParallelExecutionsPerNode', 'type': 'int'}, - 'data_flow_properties': {'key': 'dataFlowProperties', 'type': 'IntegrationRuntimeDataFlowProperties'}, - 'v_net_properties': {'key': 'vNetProperties', 'type': 'IntegrationRuntimeVNetProperties'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - location: Optional[str] = None, - node_size: Optional[str] = None, - number_of_nodes: Optional[int] = None, - max_parallel_executions_per_node: Optional[int] = None, - data_flow_properties: Optional["IntegrationRuntimeDataFlowProperties"] = None, - v_net_properties: Optional["IntegrationRuntimeVNetProperties"] = None, - **kwargs - ): - super(IntegrationRuntimeComputeProperties, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.location = location - self.node_size = node_size - self.number_of_nodes = number_of_nodes - self.max_parallel_executions_per_node = max_parallel_executions_per_node - self.data_flow_properties = data_flow_properties - self.v_net_properties = v_net_properties - - -class IntegrationRuntimeCustomSetupScriptProperties(msrest.serialization.Model): - """Custom setup script properties for a managed dedicated integration runtime. - - :param blob_container_uri: The URI of the Azure blob container that contains the custom setup - script. - :type blob_container_uri: str - :param sas_token: The SAS token of the Azure blob container. - :type sas_token: ~azure.synapse.artifacts.v2020_12_01.models.SecureString - """ - - _attribute_map = { - 'blob_container_uri': {'key': 'blobContainerUri', 'type': 'str'}, - 'sas_token': {'key': 'sasToken', 'type': 'SecureString'}, - } - - def __init__( - self, - *, - blob_container_uri: Optional[str] = None, - sas_token: Optional["SecureString"] = None, - **kwargs - ): - super(IntegrationRuntimeCustomSetupScriptProperties, self).__init__(**kwargs) - self.blob_container_uri = blob_container_uri - self.sas_token = sas_token - - -class IntegrationRuntimeDataFlowProperties(msrest.serialization.Model): - """Data flow properties for managed integration runtime. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param compute_type: Compute type of the cluster which will execute data flow job. Possible - values include: "General", "MemoryOptimized", "ComputeOptimized". - :type compute_type: str or ~azure.synapse.artifacts.v2020_12_01.models.DataFlowComputeType - :param core_count: Core count of the cluster which will execute data flow job. Supported values - are: 8, 16, 32, 48, 80, 144 and 272. - :type core_count: int - :param time_to_live: Time to live (in minutes) setting of the cluster which will execute data - flow job. - :type time_to_live: int - """ - - _validation = { - 'time_to_live': {'minimum': 0}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'compute_type': {'key': 'computeType', 'type': 'str'}, - 'core_count': {'key': 'coreCount', 'type': 'int'}, - 'time_to_live': {'key': 'timeToLive', 'type': 'int'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - compute_type: Optional[Union[str, "DataFlowComputeType"]] = None, - core_count: Optional[int] = None, - time_to_live: Optional[int] = None, - **kwargs - ): - super(IntegrationRuntimeDataFlowProperties, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.compute_type = compute_type - self.core_count = core_count - self.time_to_live = time_to_live - - -class IntegrationRuntimeDataProxyProperties(msrest.serialization.Model): - """Data proxy properties for a managed dedicated integration runtime. - - :param connect_via: The self-hosted integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.EntityReference - :param staging_linked_service: The staging linked service reference. - :type staging_linked_service: ~azure.synapse.artifacts.v2020_12_01.models.EntityReference - :param path: The path to contain the staged data in the Blob storage. - :type path: str - """ - - _attribute_map = { - 'connect_via': {'key': 'connectVia', 'type': 'EntityReference'}, - 'staging_linked_service': {'key': 'stagingLinkedService', 'type': 'EntityReference'}, - 'path': {'key': 'path', 'type': 'str'}, - } - - def __init__( - self, - *, - connect_via: Optional["EntityReference"] = None, - staging_linked_service: Optional["EntityReference"] = None, - path: Optional[str] = None, - **kwargs - ): - super(IntegrationRuntimeDataProxyProperties, self).__init__(**kwargs) - self.connect_via = connect_via - self.staging_linked_service = staging_linked_service - self.path = path - - -class IntegrationRuntimeListResponse(msrest.serialization.Model): - """A list of integration runtime resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of integration runtimes. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[IntegrationRuntimeResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - *, - value: List["IntegrationRuntimeResource"], - next_link: Optional[str] = None, - **kwargs - ): - super(IntegrationRuntimeListResponse, self).__init__(**kwargs) - self.value = value - self.next_link = next_link - - -class IntegrationRuntimeReference(msrest.serialization.Model): - """Integration runtime reference type. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Type of integration runtime. Possible values include: - "IntegrationRuntimeReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReferenceType - :param reference_name: Required. Reference integration runtime name. - :type reference_name: str - :param parameters: Arguments for integration runtime. - :type parameters: dict[str, any] - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{object}'}, - } - - def __init__( - self, - *, - type: Union[str, "IntegrationRuntimeReferenceType"], - reference_name: str, - parameters: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(IntegrationRuntimeReference, self).__init__(**kwargs) - self.type = type - self.reference_name = reference_name - self.parameters = parameters - - -class IntegrationRuntimeResource(SubResource): - """Integration runtime resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Integration runtime properties. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntime - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'IntegrationRuntime'}, - } - - def __init__( - self, - *, - properties: "IntegrationRuntime", - **kwargs - ): - super(IntegrationRuntimeResource, self).__init__(**kwargs) - self.properties = properties - - -class IntegrationRuntimeSsisCatalogInfo(msrest.serialization.Model): - """Catalog information for managed dedicated integration runtime. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param catalog_server_endpoint: The catalog database server URL. - :type catalog_server_endpoint: str - :param catalog_admin_user_name: The administrator user name of catalog database. - :type catalog_admin_user_name: str - :param catalog_admin_password: The password of the administrator user account of the catalog - database. - :type catalog_admin_password: ~azure.synapse.artifacts.v2020_12_01.models.SecureString - :param catalog_pricing_tier: The pricing tier for the catalog database. The valid values could - be found in https://azure.microsoft.com/en-us/pricing/details/sql-database/. Possible values - include: "Basic", "Standard", "Premium", "PremiumRS". - :type catalog_pricing_tier: str or - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeSsisCatalogPricingTier - """ - - _validation = { - 'catalog_admin_user_name': {'max_length': 128, 'min_length': 1}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'catalog_server_endpoint': {'key': 'catalogServerEndpoint', 'type': 'str'}, - 'catalog_admin_user_name': {'key': 'catalogAdminUserName', 'type': 'str'}, - 'catalog_admin_password': {'key': 'catalogAdminPassword', 'type': 'SecureString'}, - 'catalog_pricing_tier': {'key': 'catalogPricingTier', 'type': 'str'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - catalog_server_endpoint: Optional[str] = None, - catalog_admin_user_name: Optional[str] = None, - catalog_admin_password: Optional["SecureString"] = None, - catalog_pricing_tier: Optional[Union[str, "IntegrationRuntimeSsisCatalogPricingTier"]] = None, - **kwargs - ): - super(IntegrationRuntimeSsisCatalogInfo, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.catalog_server_endpoint = catalog_server_endpoint - self.catalog_admin_user_name = catalog_admin_user_name - self.catalog_admin_password = catalog_admin_password - self.catalog_pricing_tier = catalog_pricing_tier - - -class IntegrationRuntimeSsisProperties(msrest.serialization.Model): - """SSIS properties for managed integration runtime. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param catalog_info: Catalog information for managed dedicated integration runtime. - :type catalog_info: - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeSsisCatalogInfo - :param license_type: License type for bringing your own license scenario. Possible values - include: "BasePrice", "LicenseIncluded". - :type license_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeLicenseType - :param custom_setup_script_properties: Custom setup script properties for a managed dedicated - integration runtime. - :type custom_setup_script_properties: - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeCustomSetupScriptProperties - :param data_proxy_properties: Data proxy properties for a managed dedicated integration - runtime. - :type data_proxy_properties: - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeDataProxyProperties - :param edition: The edition for the SSIS Integration Runtime. Possible values include: - "Standard", "Enterprise". - :type edition: str or ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeEdition - :param express_custom_setup_properties: Custom setup without script properties for a SSIS - integration runtime. - :type express_custom_setup_properties: - list[~azure.synapse.artifacts.v2020_12_01.models.CustomSetupBase] - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'catalog_info': {'key': 'catalogInfo', 'type': 'IntegrationRuntimeSsisCatalogInfo'}, - 'license_type': {'key': 'licenseType', 'type': 'str'}, - 'custom_setup_script_properties': {'key': 'customSetupScriptProperties', 'type': 'IntegrationRuntimeCustomSetupScriptProperties'}, - 'data_proxy_properties': {'key': 'dataProxyProperties', 'type': 'IntegrationRuntimeDataProxyProperties'}, - 'edition': {'key': 'edition', 'type': 'str'}, - 'express_custom_setup_properties': {'key': 'expressCustomSetupProperties', 'type': '[CustomSetupBase]'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - catalog_info: Optional["IntegrationRuntimeSsisCatalogInfo"] = None, - license_type: Optional[Union[str, "IntegrationRuntimeLicenseType"]] = None, - custom_setup_script_properties: Optional["IntegrationRuntimeCustomSetupScriptProperties"] = None, - data_proxy_properties: Optional["IntegrationRuntimeDataProxyProperties"] = None, - edition: Optional[Union[str, "IntegrationRuntimeEdition"]] = None, - express_custom_setup_properties: Optional[List["CustomSetupBase"]] = None, - **kwargs - ): - super(IntegrationRuntimeSsisProperties, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.catalog_info = catalog_info - self.license_type = license_type - self.custom_setup_script_properties = custom_setup_script_properties - self.data_proxy_properties = data_proxy_properties - self.edition = edition - self.express_custom_setup_properties = express_custom_setup_properties - - -class IntegrationRuntimeVNetProperties(msrest.serialization.Model): - """VNet properties for managed integration runtime. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param v_net_id: The ID of the VNet that this integration runtime will join. - :type v_net_id: str - :param subnet: The name of the subnet this integration runtime will join. - :type subnet: str - :param public_i_ps: Resource IDs of the public IP addresses that this integration runtime will - use. - :type public_i_ps: list[str] - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'v_net_id': {'key': 'vNetId', 'type': 'str'}, - 'subnet': {'key': 'subnet', 'type': 'str'}, - 'public_i_ps': {'key': 'publicIPs', 'type': '[str]'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - v_net_id: Optional[str] = None, - subnet: Optional[str] = None, - public_i_ps: Optional[List[str]] = None, - **kwargs - ): - super(IntegrationRuntimeVNetProperties, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.v_net_id = v_net_id - self.subnet = subnet - self.public_i_ps = public_i_ps - - -class JiraLinkedService(LinkedService): - """Jira Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The IP address or host name of the Jira service. (e.g. - jira.example.com). - :type host: any - :param port: The TCP port that the Jira server uses to listen for client connections. The - default value is 443 if connecting through HTTPS, or 8080 if connecting through HTTP. - :type port: any - :param username: Required. The user name that you use to access Jira Service. - :type username: any - :param password: The password corresponding to the user name that you provided in the username - field. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'username': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - host: Any, - username: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - port: Optional[Any] = None, - password: Optional["SecretBase"] = None, - use_encrypted_endpoints: Optional[Any] = None, - use_host_verification: Optional[Any] = None, - use_peer_verification: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(JiraLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Jira' # type: str - self.host = host - self.port = port - self.username = username - self.password = password - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential - - -class JiraObjectDataset(Dataset): - """Jira Service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(JiraObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'JiraObject' # type: str - self.table_name = table_name - - -class JiraSource(TabularSource): - """A copy activity Jira Service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(JiraSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'JiraSource' # type: str - self.query = query - - -class JsonDataset(Dataset): - """Json dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param location: The location of the json data storage. - :type location: ~azure.synapse.artifacts.v2020_12_01.models.DatasetLocation - :param encoding_name: The code page name of the preferred encoding. If not specified, the - default value is UTF-8, unless BOM denotes another Unicode encoding. Refer to the name column - of the table in the following link to set supported values: - https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string (or Expression with - resultType string). - :type encoding_name: any - :param compression: The data compression method used for the json dataset. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'location': {'key': 'typeProperties.location', 'type': 'DatasetLocation'}, - 'encoding_name': {'key': 'typeProperties.encodingName', 'type': 'object'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - location: Optional["DatasetLocation"] = None, - encoding_name: Optional[Any] = None, - compression: Optional["DatasetCompression"] = None, - **kwargs - ): - super(JsonDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'Json' # type: str - self.location = location - self.encoding_name = encoding_name - self.compression = compression - - -class JsonFormat(DatasetStorageFormat): - """The data stored in JSON format. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage format.Constant filled by server. - :type type: str - :param serializer: Serializer. Type: string (or Expression with resultType string). - :type serializer: any - :param deserializer: Deserializer. Type: string (or Expression with resultType string). - :type deserializer: any - :param file_pattern: File pattern of JSON. To be more specific, the way of separating a - collection of JSON objects. The default value is 'setOfObjects'. It is case-sensitive. Possible - values include: "setOfObjects", "arrayOfObjects". - :type file_pattern: str or ~azure.synapse.artifacts.v2020_12_01.models.JsonFormatFilePattern - :param nesting_separator: The character used to separate nesting levels. Default value is '.' - (dot). Type: string (or Expression with resultType string). - :type nesting_separator: any - :param encoding_name: The code page name of the preferred encoding. If not provided, the - default value is 'utf-8', unless the byte order mark (BOM) denotes another Unicode encoding. - The full list of supported values can be found in the 'Name' column of the table of encodings - in the following reference: https://go.microsoft.com/fwlink/?linkid=861078. Type: string (or - Expression with resultType string). - :type encoding_name: any - :param json_node_reference: The JSONPath of the JSON array element to be flattened. Example: - "$.ArrayPath". Type: string (or Expression with resultType string). - :type json_node_reference: any - :param json_path_definition: The JSONPath definition for each column mapping with a customized - column name to extract data from JSON file. For fields under root object, start with "$"; for - fields inside the array chosen by jsonNodeReference property, start from the array element. - Example: {"Column1": "$.Column1Path", "Column2": "Column2PathInArray"}. Type: object (or - Expression with resultType object). - :type json_path_definition: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'serializer': {'key': 'serializer', 'type': 'object'}, - 'deserializer': {'key': 'deserializer', 'type': 'object'}, - 'file_pattern': {'key': 'filePattern', 'type': 'str'}, - 'nesting_separator': {'key': 'nestingSeparator', 'type': 'object'}, - 'encoding_name': {'key': 'encodingName', 'type': 'object'}, - 'json_node_reference': {'key': 'jsonNodeReference', 'type': 'object'}, - 'json_path_definition': {'key': 'jsonPathDefinition', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - serializer: Optional[Any] = None, - deserializer: Optional[Any] = None, - file_pattern: Optional[Union[str, "JsonFormatFilePattern"]] = None, - nesting_separator: Optional[Any] = None, - encoding_name: Optional[Any] = None, - json_node_reference: Optional[Any] = None, - json_path_definition: Optional[Any] = None, - **kwargs - ): - super(JsonFormat, self).__init__(additional_properties=additional_properties, serializer=serializer, deserializer=deserializer, **kwargs) - self.type = 'JsonFormat' # type: str - self.file_pattern = file_pattern - self.nesting_separator = nesting_separator - self.encoding_name = encoding_name - self.json_node_reference = json_node_reference - self.json_path_definition = json_path_definition - - -class JsonReadSettings(FormatReadSettings): - """Json read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param compression_properties: Compression settings. - :type compression_properties: - ~azure.synapse.artifacts.v2020_12_01.models.CompressionReadSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'compression_properties': {'key': 'compressionProperties', 'type': 'CompressionReadSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - compression_properties: Optional["CompressionReadSettings"] = None, - **kwargs - ): - super(JsonReadSettings, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'JsonReadSettings' # type: str - self.compression_properties = compression_properties - - -class JsonSink(CopySink): - """A copy activity Json sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Json store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreWriteSettings - :param format_settings: Json format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.JsonWriteSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreWriteSettings'}, - 'format_settings': {'key': 'formatSettings', 'type': 'JsonWriteSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - store_settings: Optional["StoreWriteSettings"] = None, - format_settings: Optional["JsonWriteSettings"] = None, - **kwargs - ): - super(JsonSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'JsonSink' # type: str - self.store_settings = store_settings - self.format_settings = format_settings - - -class JsonSource(CopySource): - """A copy activity Json source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Json store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - :param format_settings: Json format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.JsonReadSettings - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'format_settings': {'key': 'formatSettings', 'type': 'JsonReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - store_settings: Optional["StoreReadSettings"] = None, - format_settings: Optional["JsonReadSettings"] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(JsonSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'JsonSource' # type: str - self.store_settings = store_settings - self.format_settings = format_settings - self.additional_columns = additional_columns - - -class JsonWriteSettings(FormatWriteSettings): - """Json write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param file_pattern: File pattern of JSON. This setting controls the way a collection of JSON - objects will be treated. The default value is 'setOfObjects'. It is case-sensitive. Possible - values include: "setOfObjects", "arrayOfObjects". - :type file_pattern: str or ~azure.synapse.artifacts.v2020_12_01.models.JsonWriteFilePattern - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'file_pattern': {'key': 'filePattern', 'type': 'str'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - file_pattern: Optional[Union[str, "JsonWriteFilePattern"]] = None, - **kwargs - ): - super(JsonWriteSettings, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'JsonWriteSettings' # type: str - self.file_pattern = file_pattern - - -class LibraryInfo(msrest.serialization.Model): - """Library/package information of a Big Data pool powered by Apache Spark. - - Variables are only populated by the server, and will be ignored when sending a request. - - :param name: Name of the library. - :type name: str - :param path: Storage blob path of library. - :type path: str - :param container_name: Storage blob container name. - :type container_name: str - :ivar uploaded_timestamp: The last update time of the library. - :vartype uploaded_timestamp: ~datetime.datetime - :param type: Type of the library. - :type type: str - :ivar provisioning_status: Provisioning status of the library/package. - :vartype provisioning_status: str - :ivar creator_id: Creator Id of the library/package. - :vartype creator_id: str - """ - - _validation = { - 'uploaded_timestamp': {'readonly': True}, - 'provisioning_status': {'readonly': True}, - 'creator_id': {'readonly': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'path': {'key': 'path', 'type': 'str'}, - 'container_name': {'key': 'containerName', 'type': 'str'}, - 'uploaded_timestamp': {'key': 'uploadedTimestamp', 'type': 'iso-8601'}, - 'type': {'key': 'type', 'type': 'str'}, - 'provisioning_status': {'key': 'provisioningStatus', 'type': 'str'}, - 'creator_id': {'key': 'creatorId', 'type': 'str'}, - } - - def __init__( - self, - *, - name: Optional[str] = None, - path: Optional[str] = None, - container_name: Optional[str] = None, - type: Optional[str] = None, - **kwargs - ): - super(LibraryInfo, self).__init__(**kwargs) - self.name = name - self.path = path - self.container_name = container_name - self.uploaded_timestamp = None - self.type = type - self.provisioning_status = None - self.creator_id = None - - -class LibraryListResponse(msrest.serialization.Model): - """A list of Library resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of Library. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.LibraryResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[LibraryResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - *, - value: List["LibraryResource"], - next_link: Optional[str] = None, - **kwargs - ): - super(LibraryListResponse, self).__init__(**kwargs) - self.value = value - self.next_link = next_link - - -class LibraryRequirements(msrest.serialization.Model): - """Library requirements for a Big Data pool powered by Apache Spark. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar time: The last update time of the library requirements file. - :vartype time: ~datetime.datetime - :param content: The library requirements. - :type content: str - :param filename: The filename of the library requirements file. - :type filename: str - """ - - _validation = { - 'time': {'readonly': True}, - } - - _attribute_map = { - 'time': {'key': 'time', 'type': 'iso-8601'}, - 'content': {'key': 'content', 'type': 'str'}, - 'filename': {'key': 'filename', 'type': 'str'}, - } - - def __init__( - self, - *, - content: Optional[str] = None, - filename: Optional[str] = None, - **kwargs - ): - super(LibraryRequirements, self).__init__(**kwargs) - self.time = None - self.content = content - self.filename = filename - - -class LibraryResource(SubResource): - """Library response details. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Library/package properties. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.LibraryResourceProperties - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'LibraryResourceProperties'}, - } - - def __init__( - self, - *, - properties: "LibraryResourceProperties", - **kwargs - ): - super(LibraryResource, self).__init__(**kwargs) - self.properties = properties - - -class LibraryResourceInfo(msrest.serialization.Model): - """Library resource info. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar record_id: record Id of the library/package. - :vartype record_id: int - :ivar state: Provisioning status of the library/package. - :vartype state: str - :ivar created: The creation time of the library/package. - :vartype created: str - :ivar changed: The last updated time of the library/package. - :vartype changed: str - :ivar type: The type of the resource. E.g. LibraryArtifact. - :vartype type: str - :ivar name: Name of the library/package. - :vartype name: str - :ivar operation_id: Operation Id of the operation performed on library/package. - :vartype operation_id: str - :ivar artifact_id: artifact Id of the library/package. - :vartype artifact_id: str - """ - - _validation = { - 'id': {'readonly': True}, - 'record_id': {'readonly': True}, - 'state': {'readonly': True}, - 'created': {'readonly': True}, - 'changed': {'readonly': True}, - 'type': {'readonly': True}, - 'name': {'readonly': True}, - 'operation_id': {'readonly': True}, - 'artifact_id': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'record_id': {'key': 'recordId', 'type': 'int'}, - 'state': {'key': 'state', 'type': 'str'}, - 'created': {'key': 'created', 'type': 'str'}, - 'changed': {'key': 'changed', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'operation_id': {'key': 'operationId', 'type': 'str'}, - 'artifact_id': {'key': 'artifactId', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(LibraryResourceInfo, self).__init__(**kwargs) - self.id = None - self.record_id = None - self.state = None - self.created = None - self.changed = None - self.type = None - self.name = None - self.operation_id = None - self.artifact_id = None - - -class LibraryResourceProperties(msrest.serialization.Model): - """Library/package properties. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar name: Name of the library/package. - :vartype name: str - :ivar path: Location of library/package in storage account. - :vartype path: str - :ivar container_name: Container name of the library/package. - :vartype container_name: str - :ivar uploaded_timestamp: The last update time of the library/package. - :vartype uploaded_timestamp: str - :ivar type: Type of the library/package. - :vartype type: str - :ivar provisioning_status: Provisioning status of the library/package. - :vartype provisioning_status: str - :ivar creator_id: Creator Id of the library/package. - :vartype creator_id: str - """ - - _validation = { - 'name': {'readonly': True}, - 'path': {'readonly': True}, - 'container_name': {'readonly': True}, - 'uploaded_timestamp': {'readonly': True}, - 'type': {'readonly': True}, - 'provisioning_status': {'readonly': True}, - 'creator_id': {'readonly': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'path': {'key': 'path', 'type': 'str'}, - 'container_name': {'key': 'containerName', 'type': 'str'}, - 'uploaded_timestamp': {'key': 'uploadedTimestamp', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'provisioning_status': {'key': 'provisioningStatus', 'type': 'str'}, - 'creator_id': {'key': 'creatorId', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(LibraryResourceProperties, self).__init__(**kwargs) - self.name = None - self.path = None - self.container_name = None - self.uploaded_timestamp = None - self.type = None - self.provisioning_status = None - self.creator_id = None - - -class LinkedIntegrationRuntimeType(msrest.serialization.Model): - """The base definition of a linked integration runtime. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: LinkedIntegrationRuntimeKeyAuthorization, LinkedIntegrationRuntimeRbacAuthorization. - - All required parameters must be populated in order to send to Azure. - - :param authorization_type: Required. The authorization type for integration runtime - sharing.Constant filled by server. - :type authorization_type: str - """ - - _validation = { - 'authorization_type': {'required': True}, - } - - _attribute_map = { - 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, - } - - _subtype_map = { - 'authorization_type': {'Key': 'LinkedIntegrationRuntimeKeyAuthorization', 'RBAC': 'LinkedIntegrationRuntimeRbacAuthorization'} - } - - def __init__( - self, - **kwargs - ): - super(LinkedIntegrationRuntimeType, self).__init__(**kwargs) - self.authorization_type = None # type: Optional[str] - - -class LinkedIntegrationRuntimeKeyAuthorization(LinkedIntegrationRuntimeType): - """The key authorization type integration runtime. - - All required parameters must be populated in order to send to Azure. - - :param authorization_type: Required. The authorization type for integration runtime - sharing.Constant filled by server. - :type authorization_type: str - :param key: Required. The key used for authorization. - :type key: ~azure.synapse.artifacts.v2020_12_01.models.SecureString - """ - - _validation = { - 'authorization_type': {'required': True}, - 'key': {'required': True}, - } - - _attribute_map = { - 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, - 'key': {'key': 'key', 'type': 'SecureString'}, - } - - def __init__( - self, - *, - key: "SecureString", - **kwargs - ): - super(LinkedIntegrationRuntimeKeyAuthorization, self).__init__(**kwargs) - self.authorization_type = 'Key' # type: str - self.key = key - - -class LinkedIntegrationRuntimeRbacAuthorization(LinkedIntegrationRuntimeType): - """The role based access control (RBAC) authorization type integration runtime. - - All required parameters must be populated in order to send to Azure. - - :param authorization_type: Required. The authorization type for integration runtime - sharing.Constant filled by server. - :type authorization_type: str - :param resource_id: Required. The resource identifier of the integration runtime to be shared. - :type resource_id: str - """ - - _validation = { - 'authorization_type': {'required': True}, - 'resource_id': {'required': True}, - } - - _attribute_map = { - 'authorization_type': {'key': 'authorizationType', 'type': 'str'}, - 'resource_id': {'key': 'resourceId', 'type': 'str'}, - } - - def __init__( - self, - *, - resource_id: str, - **kwargs - ): - super(LinkedIntegrationRuntimeRbacAuthorization, self).__init__(**kwargs) - self.authorization_type = 'RBAC' # type: str - self.resource_id = resource_id - - -class LinkedServiceDebugResource(SubResourceDebugResource): - """Linked service debug resource. - - All required parameters must be populated in order to send to Azure. - - :param name: The resource name. - :type name: str - :param properties: Required. Properties of linked service. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.LinkedService - """ - - _validation = { - 'properties': {'required': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'LinkedService'}, - } - - def __init__( - self, - *, - properties: "LinkedService", - name: Optional[str] = None, - **kwargs - ): - super(LinkedServiceDebugResource, self).__init__(name=name, **kwargs) - self.properties = properties - - -class LinkedServiceListResponse(msrest.serialization.Model): - """A list of linked service resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of linked services. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[LinkedServiceResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - *, - value: List["LinkedServiceResource"], - next_link: Optional[str] = None, - **kwargs - ): - super(LinkedServiceListResponse, self).__init__(**kwargs) - self.value = value - self.next_link = next_link - - -class LinkedServiceReference(msrest.serialization.Model): - """Linked service reference type. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Linked service reference type. Possible values include: - "LinkedServiceReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.Type - :param reference_name: Required. Reference LinkedService name. - :type reference_name: str - :param parameters: Arguments for LinkedService. - :type parameters: dict[str, any] - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{object}'}, - } - - def __init__( - self, - *, - type: Union[str, "Type"], - reference_name: str, - parameters: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(LinkedServiceReference, self).__init__(**kwargs) - self.type = type - self.reference_name = reference_name - self.parameters = parameters - - -class LinkedServiceResource(SubResource): - """Linked service resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Properties of linked service. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.LinkedService - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'LinkedService'}, - } - - def __init__( - self, - *, - properties: "LinkedService", - **kwargs - ): - super(LinkedServiceResource, self).__init__(**kwargs) - self.properties = properties - - -class LogLocationSettings(msrest.serialization.Model): - """Log location settings. - - All required parameters must be populated in order to send to Azure. - - :param linked_service_name: Required. Log storage linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param path: The path to storage for storing detailed logs of activity execution. Type: string - (or Expression with resultType string). - :type path: any - """ - - _validation = { - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'path': {'key': 'path', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - path: Optional[Any] = None, - **kwargs - ): - super(LogLocationSettings, self).__init__(**kwargs) - self.linked_service_name = linked_service_name - self.path = path - - -class LogSettings(msrest.serialization.Model): - """Log settings. - - All required parameters must be populated in order to send to Azure. - - :param enable_copy_activity_log: Specifies whether to enable copy activity log. Type: boolean - (or Expression with resultType boolean). - :type enable_copy_activity_log: any - :param copy_activity_log_settings: Specifies settings for copy activity log. - :type copy_activity_log_settings: - ~azure.synapse.artifacts.v2020_12_01.models.CopyActivityLogSettings - :param log_location_settings: Required. Log location settings customer needs to provide when - enabling log. - :type log_location_settings: ~azure.synapse.artifacts.v2020_12_01.models.LogLocationSettings - """ - - _validation = { - 'log_location_settings': {'required': True}, - } - - _attribute_map = { - 'enable_copy_activity_log': {'key': 'enableCopyActivityLog', 'type': 'object'}, - 'copy_activity_log_settings': {'key': 'copyActivityLogSettings', 'type': 'CopyActivityLogSettings'}, - 'log_location_settings': {'key': 'logLocationSettings', 'type': 'LogLocationSettings'}, - } - - def __init__( - self, - *, - log_location_settings: "LogLocationSettings", - enable_copy_activity_log: Optional[Any] = None, - copy_activity_log_settings: Optional["CopyActivityLogSettings"] = None, - **kwargs - ): - super(LogSettings, self).__init__(**kwargs) - self.enable_copy_activity_log = enable_copy_activity_log - self.copy_activity_log_settings = copy_activity_log_settings - self.log_location_settings = log_location_settings - - -class LogStorageSettings(msrest.serialization.Model): - """(Deprecated. Please use LogSettings) Log storage settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param linked_service_name: Required. Log storage linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param path: The path to storage for storing detailed logs of activity execution. Type: string - (or Expression with resultType string). - :type path: any - :param log_level: Gets or sets the log level, support: Info, Warning. Type: string (or - Expression with resultType string). - :type log_level: any - :param enable_reliable_logging: Specifies whether to enable reliable logging. Type: boolean (or - Expression with resultType boolean). - :type enable_reliable_logging: any - """ - - _validation = { - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'path': {'key': 'path', 'type': 'object'}, - 'log_level': {'key': 'logLevel', 'type': 'object'}, - 'enable_reliable_logging': {'key': 'enableReliableLogging', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - path: Optional[Any] = None, - log_level: Optional[Any] = None, - enable_reliable_logging: Optional[Any] = None, - **kwargs - ): - super(LogStorageSettings, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.linked_service_name = linked_service_name - self.path = path - self.log_level = log_level - self.enable_reliable_logging = enable_reliable_logging - - -class LookupActivity(ExecutionActivity): - """Lookup activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param source: Required. Dataset-specific source properties, same as copy activity source. - :type source: ~azure.synapse.artifacts.v2020_12_01.models.CopySource - :param dataset: Required. Lookup activity dataset reference. - :type dataset: ~azure.synapse.artifacts.v2020_12_01.models.DatasetReference - :param first_row_only: Whether to return first row or all rows. Default value is true. Type: - boolean (or Expression with resultType boolean). - :type first_row_only: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'source': {'required': True}, - 'dataset': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'source': {'key': 'typeProperties.source', 'type': 'CopySource'}, - 'dataset': {'key': 'typeProperties.dataset', 'type': 'DatasetReference'}, - 'first_row_only': {'key': 'typeProperties.firstRowOnly', 'type': 'object'}, - } - - def __init__( - self, - *, - name: str, - source: "CopySource", - dataset: "DatasetReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - first_row_only: Optional[Any] = None, - **kwargs - ): - super(LookupActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'Lookup' # type: str - self.source = source - self.dataset = dataset - self.first_row_only = first_row_only - - -class MagentoLinkedService(LinkedService): - """Magento server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The URL of the Magento instance. (i.e. 192.168.222.110/magento3). - :type host: any - :param access_token: The access token from Magento. - :type access_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - host: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - access_token: Optional["SecretBase"] = None, - use_encrypted_endpoints: Optional[Any] = None, - use_host_verification: Optional[Any] = None, - use_peer_verification: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(MagentoLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Magento' # type: str - self.host = host - self.access_token = access_token - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential - - -class MagentoObjectDataset(Dataset): - """Magento server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(MagentoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'MagentoObject' # type: str - self.table_name = table_name - - -class MagentoSource(TabularSource): - """A copy activity Magento server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(MagentoSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'MagentoSource' # type: str - self.query = query - - -class ManagedIdentity(msrest.serialization.Model): - """The workspace managed identity. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar principal_id: The principal ID of the workspace managed identity. - :vartype principal_id: str - :ivar tenant_id: The tenant ID of the workspace managed identity. - :vartype tenant_id: str - :param type: The type of managed identity for the workspace. Possible values include: "None", - "SystemAssigned". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.ResourceIdentityType - """ - - _validation = { - 'principal_id': {'readonly': True}, - 'tenant_id': {'readonly': True}, - } - - _attribute_map = { - 'principal_id': {'key': 'principalId', 'type': 'str'}, - 'tenant_id': {'key': 'tenantId', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - def __init__( - self, - *, - type: Optional[Union[str, "ResourceIdentityType"]] = None, - **kwargs - ): - super(ManagedIdentity, self).__init__(**kwargs) - self.principal_id = None - self.tenant_id = None - self.type = type - - -class ManagedIntegrationRuntime(IntegrationRuntime): - """Managed integration runtime, including managed elastic and managed dedicated integration runtimes. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of integration runtime.Constant filled by server. Possible values - include: "Managed", "SelfHosted". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeType - :param description: Integration runtime description. - :type description: str - :ivar state: Integration runtime state, only valid for managed dedicated integration runtime. - Possible values include: "Initial", "Stopped", "Started", "Starting", "Stopping", - "NeedRegistration", "Online", "Limited", "Offline", "AccessDenied". - :vartype state: str or ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeState - :param managed_virtual_network: Managed Virtual Network reference. - :type managed_virtual_network: - ~azure.synapse.artifacts.v2020_12_01.models.ManagedVirtualNetworkReference - :param compute_properties: The compute resource for managed integration runtime. - :type compute_properties: - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeComputeProperties - :param ssis_properties: SSIS properties for managed integration runtime. - :type ssis_properties: - ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeSsisProperties - """ - - _validation = { - 'type': {'required': True}, - 'state': {'readonly': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'state': {'key': 'state', 'type': 'str'}, - 'managed_virtual_network': {'key': 'managedVirtualNetwork', 'type': 'ManagedVirtualNetworkReference'}, - 'compute_properties': {'key': 'typeProperties.computeProperties', 'type': 'IntegrationRuntimeComputeProperties'}, - 'ssis_properties': {'key': 'typeProperties.ssisProperties', 'type': 'IntegrationRuntimeSsisProperties'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - managed_virtual_network: Optional["ManagedVirtualNetworkReference"] = None, - compute_properties: Optional["IntegrationRuntimeComputeProperties"] = None, - ssis_properties: Optional["IntegrationRuntimeSsisProperties"] = None, - **kwargs - ): - super(ManagedIntegrationRuntime, self).__init__(additional_properties=additional_properties, description=description, **kwargs) - self.type = 'Managed' # type: str - self.state = None - self.managed_virtual_network = managed_virtual_network - self.compute_properties = compute_properties - self.ssis_properties = ssis_properties - - -class ManagedVirtualNetworkReference(msrest.serialization.Model): - """Managed Virtual Network reference type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar type: Managed Virtual Network reference type. Has constant value: - "ManagedVirtualNetworkReference". - :vartype type: str - :param reference_name: Required. Reference ManagedVirtualNetwork name. - :type reference_name: str - """ - - _validation = { - 'type': {'required': True, 'constant': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - } - - type = "ManagedVirtualNetworkReference" - - def __init__( - self, - *, - reference_name: str, - **kwargs - ): - super(ManagedVirtualNetworkReference, self).__init__(**kwargs) - self.reference_name = reference_name - - -class ManagedVirtualNetworkSettings(msrest.serialization.Model): - """Managed Virtual Network Settings. - - :param prevent_data_exfiltration: Prevent Data Exfiltration. - :type prevent_data_exfiltration: bool - :param linked_access_check_on_target_resource: Linked Access Check On Target Resource. - :type linked_access_check_on_target_resource: bool - :param allowed_aad_tenant_ids_for_linking: Allowed Aad Tenant Ids For Linking. - :type allowed_aad_tenant_ids_for_linking: list[str] - """ - - _attribute_map = { - 'prevent_data_exfiltration': {'key': 'preventDataExfiltration', 'type': 'bool'}, - 'linked_access_check_on_target_resource': {'key': 'linkedAccessCheckOnTargetResource', 'type': 'bool'}, - 'allowed_aad_tenant_ids_for_linking': {'key': 'allowedAadTenantIdsForLinking', 'type': '[str]'}, - } - - def __init__( - self, - *, - prevent_data_exfiltration: Optional[bool] = None, - linked_access_check_on_target_resource: Optional[bool] = None, - allowed_aad_tenant_ids_for_linking: Optional[List[str]] = None, - **kwargs - ): - super(ManagedVirtualNetworkSettings, self).__init__(**kwargs) - self.prevent_data_exfiltration = prevent_data_exfiltration - self.linked_access_check_on_target_resource = linked_access_check_on_target_resource - self.allowed_aad_tenant_ids_for_linking = allowed_aad_tenant_ids_for_linking - - -class MappingDataFlow(DataFlow): - """Mapping data flow. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Type of data flow.Constant filled by server. - :type type: str - :param description: The description of the data flow. - :type description: str - :param annotations: List of tags that can be used for describing the data flow. - :type annotations: list[any] - :param folder: The folder that this data flow is in. If not specified, Data flow will appear at - the root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowFolder - :param sources: List of sources in data flow. - :type sources: list[~azure.synapse.artifacts.v2020_12_01.models.DataFlowSource] - :param sinks: List of sinks in data flow. - :type sinks: list[~azure.synapse.artifacts.v2020_12_01.models.DataFlowSink] - :param transformations: List of transformations in data flow. - :type transformations: list[~azure.synapse.artifacts.v2020_12_01.models.Transformation] - :param script: DataFlow script. - :type script: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DataFlowFolder'}, - 'sources': {'key': 'typeProperties.sources', 'type': '[DataFlowSource]'}, - 'sinks': {'key': 'typeProperties.sinks', 'type': '[DataFlowSink]'}, - 'transformations': {'key': 'typeProperties.transformations', 'type': '[Transformation]'}, - 'script': {'key': 'typeProperties.script', 'type': 'str'}, - } - - def __init__( - self, - *, - description: Optional[str] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DataFlowFolder"] = None, - sources: Optional[List["DataFlowSource"]] = None, - sinks: Optional[List["DataFlowSink"]] = None, - transformations: Optional[List["Transformation"]] = None, - script: Optional[str] = None, - **kwargs - ): - super(MappingDataFlow, self).__init__(description=description, annotations=annotations, folder=folder, **kwargs) - self.type = 'MappingDataFlow' # type: str - self.sources = sources - self.sinks = sinks - self.transformations = transformations - self.script = script - - -class MariaDBLinkedService(LinkedService): - """MariaDB server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: An ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_string: Optional[Any] = None, - pwd: Optional["AzureKeyVaultSecretReference"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(MariaDBLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'MariaDB' # type: str - self.connection_string = connection_string - self.pwd = pwd - self.encrypted_credential = encrypted_credential - - -class MariaDBSource(TabularSource): - """A copy activity MariaDB server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(MariaDBSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'MariaDBSource' # type: str - self.query = query - - -class MariaDBTableDataset(Dataset): - """MariaDB server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(MariaDBTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'MariaDBTable' # type: str - self.table_name = table_name - - -class MarketoLinkedService(LinkedService): - """Marketo server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param endpoint: Required. The endpoint of the Marketo server. (i.e. 123-ABC-321.mktorest.com). - :type endpoint: any - :param client_id: Required. The client Id of your Marketo service. - :type client_id: any - :param client_secret: The client secret of your Marketo service. - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'endpoint': {'required': True}, - 'client_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - endpoint: Any, - client_id: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - client_secret: Optional["SecretBase"] = None, - use_encrypted_endpoints: Optional[Any] = None, - use_host_verification: Optional[Any] = None, - use_peer_verification: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(MarketoLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Marketo' # type: str - self.endpoint = endpoint - self.client_id = client_id - self.client_secret = client_secret - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential - - -class MarketoObjectDataset(Dataset): - """Marketo server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(MarketoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'MarketoObject' # type: str - self.table_name = table_name - - -class MarketoSource(TabularSource): - """A copy activity Marketo server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(MarketoSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'MarketoSource' # type: str - self.query = query - - -class MicrosoftAccessLinkedService(LinkedService): - """Microsoft Access linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The non-access credential portion of the connection string - as well as an optional encrypted credential. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param authentication_type: Type of authentication used to connect to the Microsoft Access as - ODBC data store. Possible values are: Anonymous and Basic. Type: string (or Expression with - resultType string). - :type authentication_type: any - :param credential: The access credential portion of the connection string specified in - driver-specific property-value format. - :type credential: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param user_name: User name for Basic authentication. Type: string (or Expression with - resultType string). - :type user_name: any - :param password: Password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, - 'credential': {'key': 'typeProperties.credential', 'type': 'SecretBase'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - connection_string: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - authentication_type: Optional[Any] = None, - credential: Optional["SecretBase"] = None, - user_name: Optional[Any] = None, - password: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(MicrosoftAccessLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'MicrosoftAccess' # type: str - self.connection_string = connection_string - self.authentication_type = authentication_type - self.credential = credential - self.user_name = user_name - self.password = password - self.encrypted_credential = encrypted_credential - - -class MicrosoftAccessSink(CopySink): - """A copy activity Microsoft Access sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param pre_copy_script: A query to execute before starting the copy. Type: string (or - Expression with resultType string). - :type pre_copy_script: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - pre_copy_script: Optional[Any] = None, - **kwargs - ): - super(MicrosoftAccessSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'MicrosoftAccessSink' # type: str - self.pre_copy_script = pre_copy_script - - -class MicrosoftAccessSource(CopySource): - """A copy activity source for Microsoft Access. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query: Optional[Any] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(MicrosoftAccessSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'MicrosoftAccessSource' # type: str - self.query = query - self.additional_columns = additional_columns - - -class MicrosoftAccessTableDataset(Dataset): - """The Microsoft Access table dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The Microsoft Access table name. Type: string (or Expression with resultType - string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(MicrosoftAccessTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'MicrosoftAccessTable' # type: str - self.table_name = table_name - - -class MongoDbAtlasCollectionDataset(Dataset): - """The MongoDB Atlas database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param collection: Required. The collection name of the MongoDB Atlas database. Type: string - (or Expression with resultType string). - :type collection: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'collection': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'collection': {'key': 'typeProperties.collection', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - collection: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - **kwargs - ): - super(MongoDbAtlasCollectionDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'MongoDbAtlasCollection' # type: str - self.collection = collection - - -class MongoDbAtlasLinkedService(LinkedService): - """Linked service for MongoDB Atlas data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The MongoDB Atlas connection string. Type: string, - SecureString or AzureKeyVaultSecretReference. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param database: Required. The name of the MongoDB Atlas database that you want to access. - Type: string (or Expression with resultType string). - :type database: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - 'database': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - } - - def __init__( - self, - *, - connection_string: Any, - database: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - **kwargs - ): - super(MongoDbAtlasLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'MongoDbAtlas' # type: str - self.connection_string = connection_string - self.database = database - - -class MongoDbAtlasSource(CopySource): - """A copy activity source for a MongoDB Atlas database. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param filter: Specifies selection filter using query operators. To return all documents in a - collection, omit this parameter or pass an empty document ({}). Type: string (or Expression - with resultType string). - :type filter: any - :param cursor_methods: Cursor methods for Mongodb query. - :type cursor_methods: - ~azure.synapse.artifacts.v2020_12_01.models.MongoDbCursorMethodsProperties - :param batch_size: Specifies the number of documents to return in each batch of the response - from MongoDB Atlas instance. In most cases, modifying the batch size will not affect the user - or the application. This property's main purpose is to avoid hit the limitation of response - size. Type: integer (or Expression with resultType integer). - :type batch_size: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'filter': {'key': 'filter', 'type': 'object'}, - 'cursor_methods': {'key': 'cursorMethods', 'type': 'MongoDbCursorMethodsProperties'}, - 'batch_size': {'key': 'batchSize', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - filter: Optional[Any] = None, - cursor_methods: Optional["MongoDbCursorMethodsProperties"] = None, - batch_size: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(MongoDbAtlasSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'MongoDbAtlasSource' # type: str - self.filter = filter - self.cursor_methods = cursor_methods - self.batch_size = batch_size - self.query_timeout = query_timeout - self.additional_columns = additional_columns - - -class MongoDbCollectionDataset(Dataset): - """The MongoDB database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param collection_name: Required. The table name of the MongoDB database. Type: string (or - Expression with resultType string). - :type collection_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'collection_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'collection_name': {'key': 'typeProperties.collectionName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - collection_name: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - **kwargs - ): - super(MongoDbCollectionDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'MongoDbCollection' # type: str - self.collection_name = collection_name - - -class MongoDbCursorMethodsProperties(msrest.serialization.Model): - """Cursor methods for Mongodb query. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param project: Specifies the fields to return in the documents that match the query filter. To - return all fields in the matching documents, omit this parameter. Type: string (or Expression - with resultType string). - :type project: any - :param sort: Specifies the order in which the query returns matching documents. Type: string - (or Expression with resultType string). Type: string (or Expression with resultType string). - :type sort: any - :param skip: Specifies the how many documents skipped and where MongoDB begins returning - results. This approach may be useful in implementing paginated results. Type: integer (or - Expression with resultType integer). - :type skip: any - :param limit: Specifies the maximum number of documents the server returns. limit() is - analogous to the LIMIT statement in a SQL database. Type: integer (or Expression with - resultType integer). - :type limit: any - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'project': {'key': 'project', 'type': 'object'}, - 'sort': {'key': 'sort', 'type': 'object'}, - 'skip': {'key': 'skip', 'type': 'object'}, - 'limit': {'key': 'limit', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - project: Optional[Any] = None, - sort: Optional[Any] = None, - skip: Optional[Any] = None, - limit: Optional[Any] = None, - **kwargs - ): - super(MongoDbCursorMethodsProperties, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.project = project - self.sort = sort - self.skip = skip - self.limit = limit - - -class MongoDbLinkedService(LinkedService): - """Linked service for MongoDb data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param server: Required. The IP address or server name of the MongoDB server. Type: string (or - Expression with resultType string). - :type server: any - :param authentication_type: The authentication type to be used to connect to the MongoDB - database. Possible values include: "Basic", "Anonymous". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.MongoDbAuthenticationType - :param database_name: Required. The name of the MongoDB database that you want to access. Type: - string (or Expression with resultType string). - :type database_name: any - :param username: Username for authentication. Type: string (or Expression with resultType - string). - :type username: any - :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param auth_source: Database to verify the username and password. Type: string (or Expression - with resultType string). - :type auth_source: any - :param port: The TCP port number that the MongoDB server uses to listen for client connections. - The default value is 27017. Type: integer (or Expression with resultType integer), minimum: 0. - :type port: any - :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The - default value is false. Type: boolean (or Expression with resultType boolean). - :type enable_ssl: any - :param allow_self_signed_server_cert: Specifies whether to allow self-signed certificates from - the server. The default value is false. Type: boolean (or Expression with resultType boolean). - :type allow_self_signed_server_cert: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'server': {'required': True}, - 'database_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'database_name': {'key': 'typeProperties.databaseName', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'auth_source': {'key': 'typeProperties.authSource', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, - 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - server: Any, - database_name: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - authentication_type: Optional[Union[str, "MongoDbAuthenticationType"]] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - auth_source: Optional[Any] = None, - port: Optional[Any] = None, - enable_ssl: Optional[Any] = None, - allow_self_signed_server_cert: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(MongoDbLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'MongoDb' # type: str - self.server = server - self.authentication_type = authentication_type - self.database_name = database_name - self.username = username - self.password = password - self.auth_source = auth_source - self.port = port - self.enable_ssl = enable_ssl - self.allow_self_signed_server_cert = allow_self_signed_server_cert - self.encrypted_credential = encrypted_credential - - -class MongoDbSource(CopySource): - """A copy activity source for a MongoDB database. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: Database query. Should be a SQL-92 query expression. Type: string (or Expression - with resultType string). - :type query: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query: Optional[Any] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(MongoDbSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'MongoDbSource' # type: str - self.query = query - self.additional_columns = additional_columns - - -class MongoDbV2CollectionDataset(Dataset): - """The MongoDB database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param collection: Required. The collection name of the MongoDB database. Type: string (or - Expression with resultType string). - :type collection: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'collection': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'collection': {'key': 'typeProperties.collection', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - collection: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - **kwargs - ): - super(MongoDbV2CollectionDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'MongoDbV2Collection' # type: str - self.collection = collection - - -class MongoDbV2LinkedService(LinkedService): - """Linked service for MongoDB data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The MongoDB connection string. Type: string, SecureString - or AzureKeyVaultSecretReference. Type: string, SecureString or AzureKeyVaultSecretReference. - :type connection_string: any - :param database: Required. The name of the MongoDB database that you want to access. Type: - string (or Expression with resultType string). - :type database: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - 'database': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - } - - def __init__( - self, - *, - connection_string: Any, - database: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - **kwargs - ): - super(MongoDbV2LinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'MongoDbV2' # type: str - self.connection_string = connection_string - self.database = database - - -class MongoDbV2Source(CopySource): - """A copy activity source for a MongoDB database. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param filter: Specifies selection filter using query operators. To return all documents in a - collection, omit this parameter or pass an empty document ({}). Type: string (or Expression - with resultType string). - :type filter: any - :param cursor_methods: Cursor methods for Mongodb query. - :type cursor_methods: - ~azure.synapse.artifacts.v2020_12_01.models.MongoDbCursorMethodsProperties - :param batch_size: Specifies the number of documents to return in each batch of the response - from MongoDB instance. In most cases, modifying the batch size will not affect the user or the - application. This property's main purpose is to avoid hit the limitation of response size. - Type: integer (or Expression with resultType integer). - :type batch_size: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'filter': {'key': 'filter', 'type': 'object'}, - 'cursor_methods': {'key': 'cursorMethods', 'type': 'MongoDbCursorMethodsProperties'}, - 'batch_size': {'key': 'batchSize', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - filter: Optional[Any] = None, - cursor_methods: Optional["MongoDbCursorMethodsProperties"] = None, - batch_size: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(MongoDbV2Source, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'MongoDbV2Source' # type: str - self.filter = filter - self.cursor_methods = cursor_methods - self.batch_size = batch_size - self.query_timeout = query_timeout - self.additional_columns = additional_columns - - -class MySqlLinkedService(LinkedService): - """Linked service for MySQL data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The connection string. - :type connection_string: any - :param password: The Azure key vault secret reference of password in connection string. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - connection_string: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - password: Optional["AzureKeyVaultSecretReference"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(MySqlLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'MySql' # type: str - self.connection_string = connection_string - self.password = password - self.encrypted_credential = encrypted_credential - - -class MySqlSource(TabularSource): - """A copy activity source for MySQL databases. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(MySqlSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'MySqlSource' # type: str - self.query = query - - -class MySqlTableDataset(Dataset): - """The MySQL table dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The MySQL table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(MySqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'MySqlTable' # type: str - self.table_name = table_name - - -class NetezzaLinkedService(LinkedService): - """Netezza linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: An ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_string: Optional[Any] = None, - pwd: Optional["AzureKeyVaultSecretReference"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(NetezzaLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Netezza' # type: str - self.connection_string = connection_string - self.pwd = pwd - self.encrypted_credential = encrypted_credential - - -class NetezzaPartitionSettings(msrest.serialization.Model): - """The settings that will be leveraged for Netezza source partitioning. - - :param partition_column_name: The name of the column in integer type that will be used for - proceeding range partitioning. Type: string (or Expression with resultType string). - :type partition_column_name: any - :param partition_upper_bound: The maximum value of column specified in partitionColumnName that - will be used for proceeding range partitioning. Type: string (or Expression with resultType - string). - :type partition_upper_bound: any - :param partition_lower_bound: The minimum value of column specified in partitionColumnName that - will be used for proceeding range partitioning. Type: string (or Expression with resultType - string). - :type partition_lower_bound: any - """ - - _attribute_map = { - 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, - 'partition_upper_bound': {'key': 'partitionUpperBound', 'type': 'object'}, - 'partition_lower_bound': {'key': 'partitionLowerBound', 'type': 'object'}, - } - - def __init__( - self, - *, - partition_column_name: Optional[Any] = None, - partition_upper_bound: Optional[Any] = None, - partition_lower_bound: Optional[Any] = None, - **kwargs - ): - super(NetezzaPartitionSettings, self).__init__(**kwargs) - self.partition_column_name = partition_column_name - self.partition_upper_bound = partition_upper_bound - self.partition_lower_bound = partition_lower_bound - - -class NetezzaSource(TabularSource): - """A copy activity Netezza source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - :param partition_option: The partition mechanism that will be used for Netezza read in - parallel. Possible values include: "None", "DataSlice", "DynamicRange". - :type partition_option: str or - ~azure.synapse.artifacts.v2020_12_01.models.NetezzaPartitionOption - :param partition_settings: The settings that will be leveraged for Netezza source partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.NetezzaPartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'NetezzaPartitionSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - partition_option: Optional[Union[str, "NetezzaPartitionOption"]] = None, - partition_settings: Optional["NetezzaPartitionSettings"] = None, - **kwargs - ): - super(NetezzaSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'NetezzaSource' # type: str - self.query = query - self.partition_option = partition_option - self.partition_settings = partition_settings - - -class NetezzaTableDataset(Dataset): - """Netezza dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The table name of the Netezza. Type: string (or Expression with resultType - string). - :type table: any - :param schema_type_properties_schema: The schema name of the Netezza. Type: string (or - Expression with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - table: Optional[Any] = None, - schema_type_properties_schema: Optional[Any] = None, - **kwargs - ): - super(NetezzaTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'NetezzaTable' # type: str - self.table_name = table_name - self.table = table - self.schema_type_properties_schema = schema_type_properties_schema - - -class Notebook(msrest.serialization.Model): - """Notebook. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param description: The description of the notebook. - :type description: str - :param big_data_pool: Big data pool reference. - :type big_data_pool: ~azure.synapse.artifacts.v2020_12_01.models.BigDataPoolReference - :param session_properties: Session properties. - :type session_properties: ~azure.synapse.artifacts.v2020_12_01.models.NotebookSessionProperties - :param metadata: Required. Notebook root-level metadata. - :type metadata: ~azure.synapse.artifacts.v2020_12_01.models.NotebookMetadata - :param nbformat: Required. Notebook format (major number). Incremented between backwards - incompatible changes to the notebook format. - :type nbformat: int - :param nbformat_minor: Required. Notebook format (minor number). Incremented for backward - compatible changes to the notebook format. - :type nbformat_minor: int - :param cells: Required. Array of cells of the current notebook. - :type cells: list[~azure.synapse.artifacts.v2020_12_01.models.NotebookCell] - """ - - _validation = { - 'metadata': {'required': True}, - 'nbformat': {'required': True}, - 'nbformat_minor': {'required': True}, - 'cells': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'description', 'type': 'str'}, - 'big_data_pool': {'key': 'bigDataPool', 'type': 'BigDataPoolReference'}, - 'session_properties': {'key': 'sessionProperties', 'type': 'NotebookSessionProperties'}, - 'metadata': {'key': 'metadata', 'type': 'NotebookMetadata'}, - 'nbformat': {'key': 'nbformat', 'type': 'int'}, - 'nbformat_minor': {'key': 'nbformat_minor', 'type': 'int'}, - 'cells': {'key': 'cells', 'type': '[NotebookCell]'}, - } - - def __init__( - self, - *, - metadata: "NotebookMetadata", - nbformat: int, - nbformat_minor: int, - cells: List["NotebookCell"], - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - big_data_pool: Optional["BigDataPoolReference"] = None, - session_properties: Optional["NotebookSessionProperties"] = None, - **kwargs - ): - super(Notebook, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.description = description - self.big_data_pool = big_data_pool - self.session_properties = session_properties - self.metadata = metadata - self.nbformat = nbformat - self.nbformat_minor = nbformat_minor - self.cells = cells - - -class NotebookCell(msrest.serialization.Model): - """Notebook cell. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param cell_type: Required. String identifying the type of cell. - :type cell_type: str - :param metadata: Required. Cell-level metadata. - :type metadata: any - :param source: Required. Contents of the cell, represented as an array of lines. - :type source: list[str] - :param attachments: Attachments associated with the cell. - :type attachments: any - :param outputs: Cell-level output items. - :type outputs: list[~azure.synapse.artifacts.v2020_12_01.models.NotebookCellOutputItem] - """ - - _validation = { - 'cell_type': {'required': True}, - 'metadata': {'required': True}, - 'source': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'cell_type': {'key': 'cell_type', 'type': 'str'}, - 'metadata': {'key': 'metadata', 'type': 'object'}, - 'source': {'key': 'source', 'type': '[str]'}, - 'attachments': {'key': 'attachments', 'type': 'object'}, - 'outputs': {'key': 'outputs', 'type': '[NotebookCellOutputItem]'}, - } - - def __init__( - self, - *, - cell_type: str, - metadata: Any, - source: List[str], - additional_properties: Optional[Dict[str, Any]] = None, - attachments: Optional[Any] = None, - outputs: Optional[List["NotebookCellOutputItem"]] = None, - **kwargs - ): - super(NotebookCell, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.cell_type = cell_type - self.metadata = metadata - self.source = source - self.attachments = attachments - self.outputs = outputs - - -class NotebookCellOutputItem(msrest.serialization.Model): - """An item of the notebook cell execution output. - - All required parameters must be populated in order to send to Azure. - - :param name: For output_type=stream, determines the name of stream (stdout / stderr). - :type name: str - :param execution_count: Execution sequence number. - :type execution_count: int - :param output_type: Required. Execution, display, or stream outputs. Possible values include: - "execute_result", "display_data", "stream", "error". - :type output_type: str or ~azure.synapse.artifacts.v2020_12_01.models.CellOutputType - :param text: For output_type=stream, the stream's text output, represented as a string or an - array of strings. - :type text: any - :param data: Output data. Use MIME type as key, and content as value. - :type data: any - :param metadata: Metadata for the output item. - :type metadata: any - """ - - _validation = { - 'output_type': {'required': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'execution_count': {'key': 'execution_count', 'type': 'int'}, - 'output_type': {'key': 'output_type', 'type': 'str'}, - 'text': {'key': 'text', 'type': 'object'}, - 'data': {'key': 'data', 'type': 'object'}, - 'metadata': {'key': 'metadata', 'type': 'object'}, - } - - def __init__( - self, - *, - output_type: Union[str, "CellOutputType"], - name: Optional[str] = None, - execution_count: Optional[int] = None, - text: Optional[Any] = None, - data: Optional[Any] = None, - metadata: Optional[Any] = None, - **kwargs - ): - super(NotebookCellOutputItem, self).__init__(**kwargs) - self.name = name - self.execution_count = execution_count - self.output_type = output_type - self.text = text - self.data = data - self.metadata = metadata - - -class NotebookKernelSpec(msrest.serialization.Model): - """Kernel information. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Name of the kernel specification. - :type name: str - :param display_name: Required. Name to display in UI. - :type display_name: str - """ - - _validation = { - 'name': {'required': True}, - 'display_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'display_name': {'key': 'display_name', 'type': 'str'}, - } - - def __init__( - self, - *, - name: str, - display_name: str, - additional_properties: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(NotebookKernelSpec, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.name = name - self.display_name = display_name - - -class NotebookLanguageInfo(msrest.serialization.Model): - """Language info. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. The programming language which this kernel runs. - :type name: str - :param codemirror_mode: The codemirror mode to use for code in this language. - :type codemirror_mode: str - """ - - _validation = { - 'name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'codemirror_mode': {'key': 'codemirror_mode', 'type': 'str'}, - } - - def __init__( - self, - *, - name: str, - additional_properties: Optional[Dict[str, Any]] = None, - codemirror_mode: Optional[str] = None, - **kwargs - ): - super(NotebookLanguageInfo, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.name = name - self.codemirror_mode = codemirror_mode - - -class NotebookListResponse(msrest.serialization.Model): - """A list of Notebook resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of Notebooks. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.NotebookResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[NotebookResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - *, - value: List["NotebookResource"], - next_link: Optional[str] = None, - **kwargs - ): - super(NotebookListResponse, self).__init__(**kwargs) - self.value = value - self.next_link = next_link - - -class NotebookMetadata(msrest.serialization.Model): - """Notebook root-level metadata. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param kernelspec: Kernel information. - :type kernelspec: ~azure.synapse.artifacts.v2020_12_01.models.NotebookKernelSpec - :param language_info: Language info. - :type language_info: ~azure.synapse.artifacts.v2020_12_01.models.NotebookLanguageInfo - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'kernelspec': {'key': 'kernelspec', 'type': 'NotebookKernelSpec'}, - 'language_info': {'key': 'language_info', 'type': 'NotebookLanguageInfo'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - kernelspec: Optional["NotebookKernelSpec"] = None, - language_info: Optional["NotebookLanguageInfo"] = None, - **kwargs - ): - super(NotebookMetadata, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.kernelspec = kernelspec - self.language_info = language_info - - -class NotebookResource(msrest.serialization.Model): - """Notebook resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource Id for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :param name: Required. The name of the resource. - :type name: str - :ivar type: The type of the resource. Ex- Microsoft.Compute/virtualMachines or - Microsoft.Storage/storageAccounts. - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Properties of Notebook. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.Notebook - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'required': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'Notebook'}, - } - - def __init__( - self, - *, - name: str, - properties: "Notebook", - **kwargs - ): - super(NotebookResource, self).__init__(**kwargs) - self.id = None - self.name = name - self.type = None - self.etag = None - self.properties = properties - - -class NotebookSessionProperties(msrest.serialization.Model): - """Session properties. - - All required parameters must be populated in order to send to Azure. - - :param driver_memory: Required. Amount of memory to use for the driver process. - :type driver_memory: str - :param driver_cores: Required. Number of cores to use for the driver. - :type driver_cores: int - :param executor_memory: Required. Amount of memory to use per executor process. - :type executor_memory: str - :param executor_cores: Required. Number of cores to use for each executor. - :type executor_cores: int - :param num_executors: Required. Number of executors to launch for this session. - :type num_executors: int - """ - - _validation = { - 'driver_memory': {'required': True}, - 'driver_cores': {'required': True}, - 'executor_memory': {'required': True}, - 'executor_cores': {'required': True}, - 'num_executors': {'required': True}, - } - - _attribute_map = { - 'driver_memory': {'key': 'driverMemory', 'type': 'str'}, - 'driver_cores': {'key': 'driverCores', 'type': 'int'}, - 'executor_memory': {'key': 'executorMemory', 'type': 'str'}, - 'executor_cores': {'key': 'executorCores', 'type': 'int'}, - 'num_executors': {'key': 'numExecutors', 'type': 'int'}, - } - - def __init__( - self, - *, - driver_memory: str, - driver_cores: int, - executor_memory: str, - executor_cores: int, - num_executors: int, - **kwargs - ): - super(NotebookSessionProperties, self).__init__(**kwargs) - self.driver_memory = driver_memory - self.driver_cores = driver_cores - self.executor_memory = executor_memory - self.executor_cores = executor_cores - self.num_executors = num_executors - - -class ODataLinkedService(LinkedService): - """Open Data Protocol (OData) linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param url: Required. The URL of the OData service endpoint. Type: string (or Expression with - resultType string). - :type url: any - :param authentication_type: Type of authentication used to connect to the OData service. - Possible values include: "Basic", "Anonymous", "Windows", "AadServicePrincipal", - "ManagedServiceIdentity". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.ODataAuthenticationType - :param user_name: User name of the OData service. Type: string (or Expression with resultType - string). - :type user_name: any - :param password: Password of the OData service. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: Specify the tenant information (domain name or tenant ID) under which your - application resides. Type: string (or Expression with resultType string). - :type tenant: any - :param service_principal_id: Specify the application id of your application registered in Azure - Active Directory. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param azure_cloud_type: Indicates the azure cloud type of the service principle auth. Allowed - values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data - factory regions’ cloud type. Type: string (or Expression with resultType string). - :type azure_cloud_type: any - :param aad_resource_id: Specify the resource you are requesting authorization to use Directory. - Type: string (or Expression with resultType string). - :type aad_resource_id: any - :param aad_service_principal_credential_type: Specify the credential type (key or cert) is used - for service principal. Possible values include: "ServicePrincipalKey", "ServicePrincipalCert". - :type aad_service_principal_credential_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.ODataAadServicePrincipalCredentialType - :param service_principal_key: Specify the secret of your application registered in Azure Active - Directory. Type: string (or Expression with resultType string). - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param service_principal_embedded_cert: Specify the base64 encoded certificate of your - application registered in Azure Active Directory. Type: string (or Expression with resultType - string). - :type service_principal_embedded_cert: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param service_principal_embedded_cert_password: Specify the password of your certificate if - your certificate has a password and you are using AadServicePrincipal authentication. Type: - string (or Expression with resultType string). - :type service_principal_embedded_cert_password: - ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'azure_cloud_type': {'key': 'typeProperties.azureCloudType', 'type': 'object'}, - 'aad_resource_id': {'key': 'typeProperties.aadResourceId', 'type': 'object'}, - 'aad_service_principal_credential_type': {'key': 'typeProperties.aadServicePrincipalCredentialType', 'type': 'str'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'service_principal_embedded_cert': {'key': 'typeProperties.servicePrincipalEmbeddedCert', 'type': 'SecretBase'}, - 'service_principal_embedded_cert_password': {'key': 'typeProperties.servicePrincipalEmbeddedCertPassword', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - url: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - authentication_type: Optional[Union[str, "ODataAuthenticationType"]] = None, - user_name: Optional[Any] = None, - password: Optional["SecretBase"] = None, - tenant: Optional[Any] = None, - service_principal_id: Optional[Any] = None, - azure_cloud_type: Optional[Any] = None, - aad_resource_id: Optional[Any] = None, - aad_service_principal_credential_type: Optional[Union[str, "ODataAadServicePrincipalCredentialType"]] = None, - service_principal_key: Optional["SecretBase"] = None, - service_principal_embedded_cert: Optional["SecretBase"] = None, - service_principal_embedded_cert_password: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(ODataLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'OData' # type: str - self.url = url - self.authentication_type = authentication_type - self.user_name = user_name - self.password = password - self.tenant = tenant - self.service_principal_id = service_principal_id - self.azure_cloud_type = azure_cloud_type - self.aad_resource_id = aad_resource_id - self.aad_service_principal_credential_type = aad_service_principal_credential_type - self.service_principal_key = service_principal_key - self.service_principal_embedded_cert = service_principal_embedded_cert - self.service_principal_embedded_cert_password = service_principal_embedded_cert_password - self.encrypted_credential = encrypted_credential - - -class ODataResourceDataset(Dataset): - """The Open Data Protocol (OData) resource dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param path: The OData resource path. Type: string (or Expression with resultType string). - :type path: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'path': {'key': 'typeProperties.path', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - path: Optional[Any] = None, - **kwargs - ): - super(ODataResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'ODataResource' # type: str - self.path = path - - -class ODataSource(CopySource): - """A copy activity source for OData source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: OData query. For example, "$top=1". Type: string (or Expression with resultType - string). - :type query: any - :param http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout - to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query: Optional[Any] = None, - http_request_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(ODataSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'ODataSource' # type: str - self.query = query - self.http_request_timeout = http_request_timeout - self.additional_columns = additional_columns - - -class OdbcLinkedService(LinkedService): - """Open Database Connectivity (ODBC) linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The non-access credential portion of the connection string - as well as an optional encrypted credential. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param authentication_type: Type of authentication used to connect to the ODBC data store. - Possible values are: Anonymous and Basic. Type: string (or Expression with resultType string). - :type authentication_type: any - :param credential: The access credential portion of the connection string specified in - driver-specific property-value format. - :type credential: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param user_name: User name for Basic authentication. Type: string (or Expression with - resultType string). - :type user_name: any - :param password: Password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, - 'credential': {'key': 'typeProperties.credential', 'type': 'SecretBase'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - connection_string: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - authentication_type: Optional[Any] = None, - credential: Optional["SecretBase"] = None, - user_name: Optional[Any] = None, - password: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(OdbcLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Odbc' # type: str - self.connection_string = connection_string - self.authentication_type = authentication_type - self.credential = credential - self.user_name = user_name - self.password = password - self.encrypted_credential = encrypted_credential - - -class OdbcSink(CopySink): - """A copy activity ODBC sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param pre_copy_script: A query to execute before starting the copy. Type: string (or - Expression with resultType string). - :type pre_copy_script: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - pre_copy_script: Optional[Any] = None, - **kwargs - ): - super(OdbcSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'OdbcSink' # type: str - self.pre_copy_script = pre_copy_script - - -class OdbcSource(TabularSource): - """A copy activity source for ODBC databases. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(OdbcSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'OdbcSource' # type: str - self.query = query - - -class OdbcTableDataset(Dataset): - """The ODBC table dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The ODBC table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(OdbcTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'OdbcTable' # type: str - self.table_name = table_name - - -class Office365Dataset(Dataset): - """The Office365 account. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: Required. Name of the dataset to extract from Office 365. Type: string (or - Expression with resultType string). - :type table_name: any - :param predicate: A predicate expression that can be used to filter the specific rows to - extract from Office 365. Type: string (or Expression with resultType string). - :type predicate: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'table_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'predicate': {'key': 'typeProperties.predicate', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - table_name: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - predicate: Optional[Any] = None, - **kwargs - ): - super(Office365Dataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'Office365Table' # type: str - self.table_name = table_name - self.predicate = predicate - - -class Office365LinkedService(LinkedService): - """Office365 linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param office365_tenant_id: Required. Azure tenant ID to which the Office 365 account belongs. - Type: string (or Expression with resultType string). - :type office365_tenant_id: any - :param service_principal_tenant_id: Required. Specify the tenant information under which your - Azure AD web application resides. Type: string (or Expression with resultType string). - :type service_principal_tenant_id: any - :param service_principal_id: Required. Specify the application's client ID. Type: string (or - Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: Required. Specify the application's key. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'office365_tenant_id': {'required': True}, - 'service_principal_tenant_id': {'required': True}, - 'service_principal_id': {'required': True}, - 'service_principal_key': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'office365_tenant_id': {'key': 'typeProperties.office365TenantId', 'type': 'object'}, - 'service_principal_tenant_id': {'key': 'typeProperties.servicePrincipalTenantId', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - office365_tenant_id: Any, - service_principal_tenant_id: Any, - service_principal_id: Any, - service_principal_key: "SecretBase", - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(Office365LinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Office365' # type: str - self.office365_tenant_id = office365_tenant_id - self.service_principal_tenant_id = service_principal_tenant_id - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.encrypted_credential = encrypted_credential - - -class Office365Source(CopySource): - """A copy activity source for an Office 365 service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param allowed_groups: The groups containing all the users. Type: array of strings (or - Expression with resultType array of strings). - :type allowed_groups: any - :param user_scope_filter_uri: The user scope uri. Type: string (or Expression with resultType - string). - :type user_scope_filter_uri: any - :param date_filter_column: The Column to apply the :code:`` and - :code:``. Type: string (or Expression with resultType string). - :type date_filter_column: any - :param start_time: Start time of the requested range for this dataset. Type: string (or - Expression with resultType string). - :type start_time: any - :param end_time: End time of the requested range for this dataset. Type: string (or Expression - with resultType string). - :type end_time: any - :param output_columns: The columns to be read out from the Office 365 table. Type: array of - objects (or Expression with resultType array of objects). Example: [ { "name": "Id" }, { - "name": "CreatedDateTime" } ]. - :type output_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'allowed_groups': {'key': 'allowedGroups', 'type': 'object'}, - 'user_scope_filter_uri': {'key': 'userScopeFilterUri', 'type': 'object'}, - 'date_filter_column': {'key': 'dateFilterColumn', 'type': 'object'}, - 'start_time': {'key': 'startTime', 'type': 'object'}, - 'end_time': {'key': 'endTime', 'type': 'object'}, - 'output_columns': {'key': 'outputColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - allowed_groups: Optional[Any] = None, - user_scope_filter_uri: Optional[Any] = None, - date_filter_column: Optional[Any] = None, - start_time: Optional[Any] = None, - end_time: Optional[Any] = None, - output_columns: Optional[Any] = None, - **kwargs - ): - super(Office365Source, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'Office365Source' # type: str - self.allowed_groups = allowed_groups - self.user_scope_filter_uri = user_scope_filter_uri - self.date_filter_column = date_filter_column - self.start_time = start_time - self.end_time = end_time - self.output_columns = output_columns - - -class OperationResult(msrest.serialization.Model): - """Operation status for the operation. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar status: Operation status. - :vartype status: str - :param code: Error code. - :type code: str - :param message: Error message. - :type message: str - :param target: Property name/path in request associated with error. - :type target: str - :param details: Array with additional error details. - :type details: list[~azure.synapse.artifacts.v2020_12_01.models.CloudError] - """ - - _validation = { - 'status': {'readonly': True}, - } - - _attribute_map = { - 'status': {'key': 'status', 'type': 'str'}, - 'code': {'key': 'error.code', 'type': 'str'}, - 'message': {'key': 'error.message', 'type': 'str'}, - 'target': {'key': 'error.target', 'type': 'str'}, - 'details': {'key': 'error.details', 'type': '[CloudError]'}, - } - - def __init__( - self, - *, - code: Optional[str] = None, - message: Optional[str] = None, - target: Optional[str] = None, - details: Optional[List["CloudError"]] = None, - **kwargs - ): - super(OperationResult, self).__init__(**kwargs) - self.status = None - self.code = code - self.message = message - self.target = target - self.details = details - - -class OracleLinkedService(LinkedService): - """Oracle database. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param password: The Azure key vault secret reference of password in connection string. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - connection_string: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - password: Optional["AzureKeyVaultSecretReference"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(OracleLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Oracle' # type: str - self.connection_string = connection_string - self.password = password - self.encrypted_credential = encrypted_credential - - -class OraclePartitionSettings(msrest.serialization.Model): - """The settings that will be leveraged for Oracle source partitioning. - - :param partition_names: Names of the physical partitions of Oracle table. - :type partition_names: any - :param partition_column_name: The name of the column in integer type that will be used for - proceeding range partitioning. Type: string (or Expression with resultType string). - :type partition_column_name: any - :param partition_upper_bound: The maximum value of column specified in partitionColumnName that - will be used for proceeding range partitioning. Type: string (or Expression with resultType - string). - :type partition_upper_bound: any - :param partition_lower_bound: The minimum value of column specified in partitionColumnName that - will be used for proceeding range partitioning. Type: string (or Expression with resultType - string). - :type partition_lower_bound: any - """ - - _attribute_map = { - 'partition_names': {'key': 'partitionNames', 'type': 'object'}, - 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, - 'partition_upper_bound': {'key': 'partitionUpperBound', 'type': 'object'}, - 'partition_lower_bound': {'key': 'partitionLowerBound', 'type': 'object'}, - } - - def __init__( - self, - *, - partition_names: Optional[Any] = None, - partition_column_name: Optional[Any] = None, - partition_upper_bound: Optional[Any] = None, - partition_lower_bound: Optional[Any] = None, - **kwargs - ): - super(OraclePartitionSettings, self).__init__(**kwargs) - self.partition_names = partition_names - self.partition_column_name = partition_column_name - self.partition_upper_bound = partition_upper_bound - self.partition_lower_bound = partition_lower_bound - - -class OracleServiceCloudLinkedService(LinkedService): - """Oracle Service Cloud linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The URL of the Oracle Service Cloud instance. - :type host: any - :param username: Required. The user name that you use to access Oracle Service Cloud server. - :type username: any - :param password: Required. The password corresponding to the user name that you provided in the - username key. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. Type: boolean (or Expression with resultType boolean). - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. Type: boolean (or Expression with resultType boolean). - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. Type: boolean (or Expression with resultType - boolean). - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'username': {'required': True}, - 'password': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - host: Any, - username: Any, - password: "SecretBase", - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - use_encrypted_endpoints: Optional[Any] = None, - use_host_verification: Optional[Any] = None, - use_peer_verification: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(OracleServiceCloudLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'OracleServiceCloud' # type: str - self.host = host - self.username = username - self.password = password - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential - - -class OracleServiceCloudObjectDataset(Dataset): - """Oracle Service Cloud dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(OracleServiceCloudObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'OracleServiceCloudObject' # type: str - self.table_name = table_name - - -class OracleServiceCloudSource(TabularSource): - """A copy activity Oracle Service Cloud source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(OracleServiceCloudSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'OracleServiceCloudSource' # type: str - self.query = query - - -class OracleSink(CopySink): - """A copy activity Oracle sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType - string). - :type pre_copy_script: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - pre_copy_script: Optional[Any] = None, - **kwargs - ): - super(OracleSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'OracleSink' # type: str - self.pre_copy_script = pre_copy_script - - -class OracleSource(CopySource): - """A copy activity Oracle source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param oracle_reader_query: Oracle reader query. Type: string (or Expression with resultType - string). - :type oracle_reader_query: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param partition_option: The partition mechanism that will be used for Oracle read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". - :type partition_option: str or - ~azure.synapse.artifacts.v2020_12_01.models.OraclePartitionOption - :param partition_settings: The settings that will be leveraged for Oracle source partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.OraclePartitionSettings - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'oracle_reader_query': {'key': 'oracleReaderQuery', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'OraclePartitionSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - oracle_reader_query: Optional[Any] = None, - query_timeout: Optional[Any] = None, - partition_option: Optional[Union[str, "OraclePartitionOption"]] = None, - partition_settings: Optional["OraclePartitionSettings"] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(OracleSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'OracleSource' # type: str - self.oracle_reader_query = oracle_reader_query - self.query_timeout = query_timeout - self.partition_option = partition_option - self.partition_settings = partition_settings - self.additional_columns = additional_columns - - -class OracleTableDataset(Dataset): - """The on-premises Oracle database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param schema_type_properties_schema: The schema name of the on-premises Oracle database. Type: - string (or Expression with resultType string). - :type schema_type_properties_schema: any - :param table: The table name of the on-premises Oracle database. Type: string (or Expression - with resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - schema_type_properties_schema: Optional[Any] = None, - table: Optional[Any] = None, - **kwargs - ): - super(OracleTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'OracleTable' # type: str - self.table_name = table_name - self.schema_type_properties_schema = schema_type_properties_schema - self.table = table - - -class OrcDataset(Dataset): - """ORC dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param location: The location of the ORC data storage. - :type location: ~azure.synapse.artifacts.v2020_12_01.models.DatasetLocation - :param orc_compression_codec: Possible values include: "none", "zlib", "snappy", "lzo". - :type orc_compression_codec: str or - ~azure.synapse.artifacts.v2020_12_01.models.OrcCompressionCodec - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'location': {'key': 'typeProperties.location', 'type': 'DatasetLocation'}, - 'orc_compression_codec': {'key': 'typeProperties.orcCompressionCodec', 'type': 'str'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - location: Optional["DatasetLocation"] = None, - orc_compression_codec: Optional[Union[str, "OrcCompressionCodec"]] = None, - **kwargs - ): - super(OrcDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'Orc' # type: str - self.location = location - self.orc_compression_codec = orc_compression_codec - - -class OrcFormat(DatasetStorageFormat): - """The data stored in Optimized Row Columnar (ORC) format. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage format.Constant filled by server. - :type type: str - :param serializer: Serializer. Type: string (or Expression with resultType string). - :type serializer: any - :param deserializer: Deserializer. Type: string (or Expression with resultType string). - :type deserializer: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'serializer': {'key': 'serializer', 'type': 'object'}, - 'deserializer': {'key': 'deserializer', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - serializer: Optional[Any] = None, - deserializer: Optional[Any] = None, - **kwargs - ): - super(OrcFormat, self).__init__(additional_properties=additional_properties, serializer=serializer, deserializer=deserializer, **kwargs) - self.type = 'OrcFormat' # type: str - - -class OrcSink(CopySink): - """A copy activity ORC sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: ORC store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreWriteSettings - :param format_settings: ORC format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.OrcWriteSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreWriteSettings'}, - 'format_settings': {'key': 'formatSettings', 'type': 'OrcWriteSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - store_settings: Optional["StoreWriteSettings"] = None, - format_settings: Optional["OrcWriteSettings"] = None, - **kwargs - ): - super(OrcSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'OrcSink' # type: str - self.store_settings = store_settings - self.format_settings = format_settings - - -class OrcSource(CopySource): - """A copy activity ORC source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: ORC store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - store_settings: Optional["StoreReadSettings"] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(OrcSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'OrcSource' # type: str - self.store_settings = store_settings - self.additional_columns = additional_columns - - -class OrcWriteSettings(FormatWriteSettings): - """Orc write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param max_rows_per_file: Limit the written file's row count to be smaller than or equal to the - specified count. Type: integer (or Expression with resultType integer). - :type max_rows_per_file: any - :param file_name_prefix: Specifies the file name pattern - :code:``_:code:``.:code:`` when copy from non-file - based store without partitionOptions. Type: string (or Expression with resultType string). - :type file_name_prefix: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_rows_per_file': {'key': 'maxRowsPerFile', 'type': 'object'}, - 'file_name_prefix': {'key': 'fileNamePrefix', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_rows_per_file: Optional[Any] = None, - file_name_prefix: Optional[Any] = None, - **kwargs - ): - super(OrcWriteSettings, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'OrcWriteSettings' # type: str - self.max_rows_per_file = max_rows_per_file - self.file_name_prefix = file_name_prefix - - -class ParameterSpecification(msrest.serialization.Model): - """Definition of a single parameter for an entity. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Parameter type. Possible values include: "Object", "String", "Int", - "Float", "Bool", "Array", "SecureString". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.ParameterType - :param default_value: Default value of parameter. - :type default_value: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'default_value': {'key': 'defaultValue', 'type': 'object'}, - } - - def __init__( - self, - *, - type: Union[str, "ParameterType"], - default_value: Optional[Any] = None, - **kwargs - ): - super(ParameterSpecification, self).__init__(**kwargs) - self.type = type - self.default_value = default_value - - -class ParquetDataset(Dataset): - """Parquet dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param location: The location of the parquet storage. - :type location: ~azure.synapse.artifacts.v2020_12_01.models.DatasetLocation - :param compression_codec: A string from ParquetCompressionCodecEnum or an expression. - :type compression_codec: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'location': {'key': 'typeProperties.location', 'type': 'DatasetLocation'}, - 'compression_codec': {'key': 'typeProperties.compressionCodec', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - location: Optional["DatasetLocation"] = None, - compression_codec: Optional[Any] = None, - **kwargs - ): - super(ParquetDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'Parquet' # type: str - self.location = location - self.compression_codec = compression_codec - - -class ParquetFormat(DatasetStorageFormat): - """The data stored in Parquet format. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage format.Constant filled by server. - :type type: str - :param serializer: Serializer. Type: string (or Expression with resultType string). - :type serializer: any - :param deserializer: Deserializer. Type: string (or Expression with resultType string). - :type deserializer: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'serializer': {'key': 'serializer', 'type': 'object'}, - 'deserializer': {'key': 'deserializer', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - serializer: Optional[Any] = None, - deserializer: Optional[Any] = None, - **kwargs - ): - super(ParquetFormat, self).__init__(additional_properties=additional_properties, serializer=serializer, deserializer=deserializer, **kwargs) - self.type = 'ParquetFormat' # type: str - - -class ParquetSink(CopySink): - """A copy activity Parquet sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Parquet store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreWriteSettings - :param format_settings: Parquet format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.ParquetWriteSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreWriteSettings'}, - 'format_settings': {'key': 'formatSettings', 'type': 'ParquetWriteSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - store_settings: Optional["StoreWriteSettings"] = None, - format_settings: Optional["ParquetWriteSettings"] = None, - **kwargs - ): - super(ParquetSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'ParquetSink' # type: str - self.store_settings = store_settings - self.format_settings = format_settings - - -class ParquetSource(CopySource): - """A copy activity Parquet source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Parquet store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - store_settings: Optional["StoreReadSettings"] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(ParquetSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'ParquetSource' # type: str - self.store_settings = store_settings - self.additional_columns = additional_columns - - -class ParquetWriteSettings(FormatWriteSettings): - """Parquet write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param max_rows_per_file: Limit the written file's row count to be smaller than or equal to the - specified count. Type: integer (or Expression with resultType integer). - :type max_rows_per_file: any - :param file_name_prefix: Specifies the file name pattern - :code:``_:code:``.:code:`` when copy from non-file - based store without partitionOptions. Type: string (or Expression with resultType string). - :type file_name_prefix: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_rows_per_file': {'key': 'maxRowsPerFile', 'type': 'object'}, - 'file_name_prefix': {'key': 'fileNamePrefix', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_rows_per_file: Optional[Any] = None, - file_name_prefix: Optional[Any] = None, - **kwargs - ): - super(ParquetWriteSettings, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'ParquetWriteSettings' # type: str - self.max_rows_per_file = max_rows_per_file - self.file_name_prefix = file_name_prefix - - -class PaypalLinkedService(LinkedService): - """Paypal Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The URL of the PayPal instance. (i.e. api.sandbox.paypal.com). - :type host: any - :param client_id: Required. The client ID associated with your PayPal application. - :type client_id: any - :param client_secret: The client secret associated with your PayPal application. - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'client_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - host: Any, - client_id: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - client_secret: Optional["SecretBase"] = None, - use_encrypted_endpoints: Optional[Any] = None, - use_host_verification: Optional[Any] = None, - use_peer_verification: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(PaypalLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Paypal' # type: str - self.host = host - self.client_id = client_id - self.client_secret = client_secret - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential - - -class PaypalObjectDataset(Dataset): - """Paypal Service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(PaypalObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'PaypalObject' # type: str - self.table_name = table_name - - -class PaypalSource(TabularSource): - """A copy activity Paypal Service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(PaypalSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'PaypalSource' # type: str - self.query = query - - -class PhoenixLinkedService(LinkedService): - """Phoenix server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The IP address or host name of the Phoenix server. (i.e. - 192.168.222.160). - :type host: any - :param port: The TCP port that the Phoenix server uses to listen for client connections. The - default value is 8765. - :type port: any - :param http_path: The partial URL corresponding to the Phoenix server. (i.e. - /gateway/sandbox/phoenix/version). The default value is hbasephoenix if using - WindowsAzureHDInsightService. - :type http_path: any - :param authentication_type: Required. The authentication mechanism used to connect to the - Phoenix server. Possible values include: "Anonymous", "UsernameAndPassword", - "WindowsAzureHDInsightService". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.PhoenixAuthenticationType - :param username: The user name used to connect to the Phoenix server. - :type username: any - :param password: The password corresponding to the user name. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The - default value is false. - :type enable_ssl: any - :param trusted_cert_path: The full path of the .pem file containing trusted CA certificates for - verifying the server when connecting over SSL. This property can only be set when using SSL on - self-hosted IR. The default value is the cacerts.pem file installed with the IR. - :type trusted_cert_path: any - :param use_system_trust_store: Specifies whether to use a CA certificate from the system trust - store or from a specified PEM file. The default value is false. - :type use_system_trust_store: any - :param allow_host_name_cn_mismatch: Specifies whether to require a CA-issued SSL certificate - name to match the host name of the server when connecting over SSL. The default value is false. - :type allow_host_name_cn_mismatch: any - :param allow_self_signed_server_cert: Specifies whether to allow self-signed certificates from - the server. The default value is false. - :type allow_self_signed_server_cert: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'http_path': {'key': 'typeProperties.httpPath', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, - 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, - 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, - 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, - 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - host: Any, - authentication_type: Union[str, "PhoenixAuthenticationType"], - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - port: Optional[Any] = None, - http_path: Optional[Any] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - enable_ssl: Optional[Any] = None, - trusted_cert_path: Optional[Any] = None, - use_system_trust_store: Optional[Any] = None, - allow_host_name_cn_mismatch: Optional[Any] = None, - allow_self_signed_server_cert: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(PhoenixLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Phoenix' # type: str - self.host = host - self.port = port - self.http_path = http_path - self.authentication_type = authentication_type - self.username = username - self.password = password - self.enable_ssl = enable_ssl - self.trusted_cert_path = trusted_cert_path - self.use_system_trust_store = use_system_trust_store - self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch - self.allow_self_signed_server_cert = allow_self_signed_server_cert - self.encrypted_credential = encrypted_credential - - -class PhoenixObjectDataset(Dataset): - """Phoenix server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The table name of the Phoenix. Type: string (or Expression with resultType - string). - :type table: any - :param schema_type_properties_schema: The schema name of the Phoenix. Type: string (or - Expression with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - table: Optional[Any] = None, - schema_type_properties_schema: Optional[Any] = None, - **kwargs - ): - super(PhoenixObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'PhoenixObject' # type: str - self.table_name = table_name - self.table = table - self.schema_type_properties_schema = schema_type_properties_schema - - -class PhoenixSource(TabularSource): - """A copy activity Phoenix server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(PhoenixSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'PhoenixSource' # type: str - self.query = query - - -class PipelineFolder(msrest.serialization.Model): - """The folder that this Pipeline is in. If not specified, Pipeline will appear at the root level. - - :param name: The name of the folder that this Pipeline is in. - :type name: str - """ - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - } - - def __init__( - self, - *, - name: Optional[str] = None, - **kwargs - ): - super(PipelineFolder, self).__init__(**kwargs) - self.name = name - - -class PipelineListResponse(msrest.serialization.Model): - """A list of pipeline resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of pipelines. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.PipelineResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[PipelineResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - *, - value: List["PipelineResource"], - next_link: Optional[str] = None, - **kwargs - ): - super(PipelineListResponse, self).__init__(**kwargs) - self.value = value - self.next_link = next_link - - -class PipelineReference(msrest.serialization.Model): - """Pipeline reference type. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Pipeline reference type. Possible values include: "PipelineReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.PipelineReferenceType - :param reference_name: Required. Reference pipeline name. - :type reference_name: str - :param name: Reference name. - :type name: str - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - } - - def __init__( - self, - *, - type: Union[str, "PipelineReferenceType"], - reference_name: str, - name: Optional[str] = None, - **kwargs - ): - super(PipelineReference, self).__init__(**kwargs) - self.type = type - self.reference_name = reference_name - self.name = name - - -class PipelineResource(SubResource): - """Pipeline resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param description: The description of the pipeline. - :type description: str - :param activities: List of activities in pipeline. - :type activities: list[~azure.synapse.artifacts.v2020_12_01.models.Activity] - :param parameters: List of parameters for pipeline. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param variables: List of variables for pipeline. - :type variables: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.VariableSpecification] - :param concurrency: The max number of concurrent runs for the pipeline. - :type concurrency: int - :param annotations: List of tags that can be used for describing the Pipeline. - :type annotations: list[any] - :param run_dimensions: Dimensions emitted by Pipeline. - :type run_dimensions: dict[str, any] - :param folder: The folder that this Pipeline is in. If not specified, Pipeline will appear at - the root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.PipelineFolder - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'concurrency': {'minimum': 1}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'properties.description', 'type': 'str'}, - 'activities': {'key': 'properties.activities', 'type': '[Activity]'}, - 'parameters': {'key': 'properties.parameters', 'type': '{ParameterSpecification}'}, - 'variables': {'key': 'properties.variables', 'type': '{VariableSpecification}'}, - 'concurrency': {'key': 'properties.concurrency', 'type': 'int'}, - 'annotations': {'key': 'properties.annotations', 'type': '[object]'}, - 'run_dimensions': {'key': 'properties.runDimensions', 'type': '{object}'}, - 'folder': {'key': 'properties.folder', 'type': 'PipelineFolder'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - activities: Optional[List["Activity"]] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - variables: Optional[Dict[str, "VariableSpecification"]] = None, - concurrency: Optional[int] = None, - annotations: Optional[List[Any]] = None, - run_dimensions: Optional[Dict[str, Any]] = None, - folder: Optional["PipelineFolder"] = None, - **kwargs - ): - super(PipelineResource, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.description = description - self.activities = activities - self.parameters = parameters - self.variables = variables - self.concurrency = concurrency - self.annotations = annotations - self.run_dimensions = run_dimensions - self.folder = folder - - -class PipelineRun(msrest.serialization.Model): - """Information about a pipeline run. - - Variables are only populated by the server, and will be ignored when sending a request. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :ivar run_id: Identifier of a run. - :vartype run_id: str - :ivar run_group_id: Identifier that correlates all the recovery runs of a pipeline run. - :vartype run_group_id: str - :ivar is_latest: Indicates if the recovered pipeline run is the latest in its group. - :vartype is_latest: bool - :ivar pipeline_name: The pipeline name. - :vartype pipeline_name: str - :ivar parameters: The full or partial list of parameter name, value pair used in the pipeline - run. - :vartype parameters: dict[str, str] - :ivar invoked_by: Entity that started the pipeline run. - :vartype invoked_by: ~azure.synapse.artifacts.v2020_12_01.models.PipelineRunInvokedBy - :ivar last_updated: The last updated timestamp for the pipeline run event in ISO8601 format. - :vartype last_updated: ~datetime.datetime - :ivar run_start: The start time of a pipeline run in ISO8601 format. - :vartype run_start: ~datetime.datetime - :ivar run_end: The end time of a pipeline run in ISO8601 format. - :vartype run_end: ~datetime.datetime - :ivar duration_in_ms: The duration of a pipeline run. - :vartype duration_in_ms: int - :ivar status: The status of a pipeline run. - :vartype status: str - :ivar message: The message from a pipeline run. - :vartype message: str - """ - - _validation = { - 'run_id': {'readonly': True}, - 'run_group_id': {'readonly': True}, - 'is_latest': {'readonly': True}, - 'pipeline_name': {'readonly': True}, - 'parameters': {'readonly': True}, - 'invoked_by': {'readonly': True}, - 'last_updated': {'readonly': True}, - 'run_start': {'readonly': True}, - 'run_end': {'readonly': True}, - 'duration_in_ms': {'readonly': True}, - 'status': {'readonly': True}, - 'message': {'readonly': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'run_id': {'key': 'runId', 'type': 'str'}, - 'run_group_id': {'key': 'runGroupId', 'type': 'str'}, - 'is_latest': {'key': 'isLatest', 'type': 'bool'}, - 'pipeline_name': {'key': 'pipelineName', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{str}'}, - 'invoked_by': {'key': 'invokedBy', 'type': 'PipelineRunInvokedBy'}, - 'last_updated': {'key': 'lastUpdated', 'type': 'iso-8601'}, - 'run_start': {'key': 'runStart', 'type': 'iso-8601'}, - 'run_end': {'key': 'runEnd', 'type': 'iso-8601'}, - 'duration_in_ms': {'key': 'durationInMs', 'type': 'int'}, - 'status': {'key': 'status', 'type': 'str'}, - 'message': {'key': 'message', 'type': 'str'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(PipelineRun, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.run_id = None - self.run_group_id = None - self.is_latest = None - self.pipeline_name = None - self.parameters = None - self.invoked_by = None - self.last_updated = None - self.run_start = None - self.run_end = None - self.duration_in_ms = None - self.status = None - self.message = None - - -class PipelineRunInvokedBy(msrest.serialization.Model): - """Provides entity name and id that started the pipeline run. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar name: Name of the entity that started the pipeline run. - :vartype name: str - :ivar id: The ID of the entity that started the run. - :vartype id: str - :ivar invoked_by_type: The type of the entity that started the run. - :vartype invoked_by_type: str - """ - - _validation = { - 'name': {'readonly': True}, - 'id': {'readonly': True}, - 'invoked_by_type': {'readonly': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'id': {'key': 'id', 'type': 'str'}, - 'invoked_by_type': {'key': 'invokedByType', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(PipelineRunInvokedBy, self).__init__(**kwargs) - self.name = None - self.id = None - self.invoked_by_type = None - - -class PipelineRunsQueryResponse(msrest.serialization.Model): - """A list pipeline runs. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of pipeline runs. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.PipelineRun] - :param continuation_token: The continuation token for getting the next page of results, if any - remaining results exist, null otherwise. - :type continuation_token: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[PipelineRun]'}, - 'continuation_token': {'key': 'continuationToken', 'type': 'str'}, - } - - def __init__( - self, - *, - value: List["PipelineRun"], - continuation_token: Optional[str] = None, - **kwargs - ): - super(PipelineRunsQueryResponse, self).__init__(**kwargs) - self.value = value - self.continuation_token = continuation_token - - -class PolybaseSettings(msrest.serialization.Model): - """PolyBase settings. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param reject_type: Reject type. Possible values include: "value", "percentage". - :type reject_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.PolybaseSettingsRejectType - :param reject_value: Specifies the value or the percentage of rows that can be rejected before - the query fails. Type: number (or Expression with resultType number), minimum: 0. - :type reject_value: any - :param reject_sample_value: Determines the number of rows to attempt to retrieve before the - PolyBase recalculates the percentage of rejected rows. Type: integer (or Expression with - resultType integer), minimum: 0. - :type reject_sample_value: any - :param use_type_default: Specifies how to handle missing values in delimited text files when - PolyBase retrieves data from the text file. Type: boolean (or Expression with resultType - boolean). - :type use_type_default: any - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'reject_type': {'key': 'rejectType', 'type': 'str'}, - 'reject_value': {'key': 'rejectValue', 'type': 'object'}, - 'reject_sample_value': {'key': 'rejectSampleValue', 'type': 'object'}, - 'use_type_default': {'key': 'useTypeDefault', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - reject_type: Optional[Union[str, "PolybaseSettingsRejectType"]] = None, - reject_value: Optional[Any] = None, - reject_sample_value: Optional[Any] = None, - use_type_default: Optional[Any] = None, - **kwargs - ): - super(PolybaseSettings, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.reject_type = reject_type - self.reject_value = reject_value - self.reject_sample_value = reject_sample_value - self.use_type_default = use_type_default - - -class PostgreSqlLinkedService(LinkedService): - """Linked service for PostgreSQL data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The connection string. - :type connection_string: any - :param password: The Azure key vault secret reference of password in connection string. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - connection_string: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - password: Optional["AzureKeyVaultSecretReference"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(PostgreSqlLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'PostgreSql' # type: str - self.connection_string = connection_string - self.password = password - self.encrypted_credential = encrypted_credential - - -class PostgreSqlSource(TabularSource): - """A copy activity source for PostgreSQL databases. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(PostgreSqlSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'PostgreSqlSource' # type: str - self.query = query - - -class PostgreSqlTableDataset(Dataset): - """The PostgreSQL table dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The PostgreSQL table name. Type: string (or Expression with resultType string). - :type table: any - :param schema_type_properties_schema: The PostgreSQL schema name. Type: string (or Expression - with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - table: Optional[Any] = None, - schema_type_properties_schema: Optional[Any] = None, - **kwargs - ): - super(PostgreSqlTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'PostgreSqlTable' # type: str - self.table_name = table_name - self.table = table - self.schema_type_properties_schema = schema_type_properties_schema - - -class PrestoLinkedService(LinkedService): - """Presto server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The IP address or host name of the Presto server. (i.e. - 192.168.222.160). - :type host: any - :param server_version: Required. The version of the Presto server. (i.e. 0.148-t). - :type server_version: any - :param catalog: Required. The catalog context for all request against the server. - :type catalog: any - :param port: The TCP port that the Presto server uses to listen for client connections. The - default value is 8080. - :type port: any - :param authentication_type: Required. The authentication mechanism used to connect to the - Presto server. Possible values include: "Anonymous", "LDAP". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.PrestoAuthenticationType - :param username: The user name used to connect to the Presto server. - :type username: any - :param password: The password corresponding to the user name. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The - default value is false. - :type enable_ssl: any - :param trusted_cert_path: The full path of the .pem file containing trusted CA certificates for - verifying the server when connecting over SSL. This property can only be set when using SSL on - self-hosted IR. The default value is the cacerts.pem file installed with the IR. - :type trusted_cert_path: any - :param use_system_trust_store: Specifies whether to use a CA certificate from the system trust - store or from a specified PEM file. The default value is false. - :type use_system_trust_store: any - :param allow_host_name_cn_mismatch: Specifies whether to require a CA-issued SSL certificate - name to match the host name of the server when connecting over SSL. The default value is false. - :type allow_host_name_cn_mismatch: any - :param allow_self_signed_server_cert: Specifies whether to allow self-signed certificates from - the server. The default value is false. - :type allow_self_signed_server_cert: any - :param time_zone_id: The local time zone used by the connection. Valid values for this option - are specified in the IANA Time Zone Database. The default value is the system time zone. - :type time_zone_id: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'server_version': {'required': True}, - 'catalog': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'server_version': {'key': 'typeProperties.serverVersion', 'type': 'object'}, - 'catalog': {'key': 'typeProperties.catalog', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, - 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, - 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, - 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, - 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, - 'time_zone_id': {'key': 'typeProperties.timeZoneID', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - host: Any, - server_version: Any, - catalog: Any, - authentication_type: Union[str, "PrestoAuthenticationType"], - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - port: Optional[Any] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - enable_ssl: Optional[Any] = None, - trusted_cert_path: Optional[Any] = None, - use_system_trust_store: Optional[Any] = None, - allow_host_name_cn_mismatch: Optional[Any] = None, - allow_self_signed_server_cert: Optional[Any] = None, - time_zone_id: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(PrestoLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Presto' # type: str - self.host = host - self.server_version = server_version - self.catalog = catalog - self.port = port - self.authentication_type = authentication_type - self.username = username - self.password = password - self.enable_ssl = enable_ssl - self.trusted_cert_path = trusted_cert_path - self.use_system_trust_store = use_system_trust_store - self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch - self.allow_self_signed_server_cert = allow_self_signed_server_cert - self.time_zone_id = time_zone_id - self.encrypted_credential = encrypted_credential - - -class PrestoObjectDataset(Dataset): - """Presto server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The table name of the Presto. Type: string (or Expression with resultType - string). - :type table: any - :param schema_type_properties_schema: The schema name of the Presto. Type: string (or - Expression with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - table: Optional[Any] = None, - schema_type_properties_schema: Optional[Any] = None, - **kwargs - ): - super(PrestoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'PrestoObject' # type: str - self.table_name = table_name - self.table = table - self.schema_type_properties_schema = schema_type_properties_schema - - -class PrestoSource(TabularSource): - """A copy activity Presto server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(PrestoSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'PrestoSource' # type: str - self.query = query - - -class PrivateEndpoint(msrest.serialization.Model): - """Private endpoint details. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar id: Resource id of the private endpoint. - :vartype id: str - """ - - _validation = { - 'id': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(PrivateEndpoint, self).__init__(**kwargs) - self.id = None - - -class ProxyResource(Resource): - """The resource model definition for a Azure Resource Manager proxy resource. It will not have tags and a location. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(ProxyResource, self).__init__(**kwargs) - - -class PrivateEndpointConnection(ProxyResource): - """A private endpoint connection. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :param private_endpoint: The private endpoint which the connection belongs to. - :type private_endpoint: ~azure.synapse.artifacts.v2020_12_01.models.PrivateEndpoint - :param private_link_service_connection_state: Connection state of the private endpoint - connection. - :type private_link_service_connection_state: - ~azure.synapse.artifacts.v2020_12_01.models.PrivateLinkServiceConnectionState - :ivar provisioning_state: Provisioning state of the private endpoint connection. - :vartype provisioning_state: str - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'provisioning_state': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'private_endpoint': {'key': 'properties.privateEndpoint', 'type': 'PrivateEndpoint'}, - 'private_link_service_connection_state': {'key': 'properties.privateLinkServiceConnectionState', 'type': 'PrivateLinkServiceConnectionState'}, - 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, - } - - def __init__( - self, - *, - private_endpoint: Optional["PrivateEndpoint"] = None, - private_link_service_connection_state: Optional["PrivateLinkServiceConnectionState"] = None, - **kwargs - ): - super(PrivateEndpointConnection, self).__init__(**kwargs) - self.private_endpoint = private_endpoint - self.private_link_service_connection_state = private_link_service_connection_state - self.provisioning_state = None - - -class PrivateLinkServiceConnectionState(msrest.serialization.Model): - """Connection state details of the private endpoint. - - Variables are only populated by the server, and will be ignored when sending a request. - - :param status: The private link service connection status. - :type status: str - :param description: The private link service connection description. - :type description: str - :ivar actions_required: The actions required for private link service connection. - :vartype actions_required: str - """ - - _validation = { - 'actions_required': {'readonly': True}, - } - - _attribute_map = { - 'status': {'key': 'status', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'actions_required': {'key': 'actionsRequired', 'type': 'str'}, - } - - def __init__( - self, - *, - status: Optional[str] = None, - description: Optional[str] = None, - **kwargs - ): - super(PrivateLinkServiceConnectionState, self).__init__(**kwargs) - self.status = status - self.description = description - self.actions_required = None - - -class PurviewConfiguration(msrest.serialization.Model): - """Purview Configuration. - - :param purview_resource_id: Purview Resource ID. - :type purview_resource_id: str - """ - - _attribute_map = { - 'purview_resource_id': {'key': 'purviewResourceId', 'type': 'str'}, - } - - def __init__( - self, - *, - purview_resource_id: Optional[str] = None, - **kwargs - ): - super(PurviewConfiguration, self).__init__(**kwargs) - self.purview_resource_id = purview_resource_id - - -class QueryDataFlowDebugSessionsResponse(msrest.serialization.Model): - """A list of active debug sessions. - - :param value: Array with all active debug sessions. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.DataFlowDebugSessionInfo] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _attribute_map = { - 'value': {'key': 'value', 'type': '[DataFlowDebugSessionInfo]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - *, - value: Optional[List["DataFlowDebugSessionInfo"]] = None, - next_link: Optional[str] = None, - **kwargs - ): - super(QueryDataFlowDebugSessionsResponse, self).__init__(**kwargs) - self.value = value - self.next_link = next_link - - -class QuickBooksLinkedService(LinkedService): - """QuickBooks server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_properties: Properties used to connect to QuickBooks. It is mutually - exclusive with any other properties in the linked service. Type: object. - :type connection_properties: any - :param endpoint: Required. The endpoint of the QuickBooks server. (i.e. - quickbooks.api.intuit.com). - :type endpoint: any - :param company_id: Required. The company ID of the QuickBooks company to authorize. - :type company_id: any - :param consumer_key: Required. The consumer key for OAuth 1.0 authentication. - :type consumer_key: any - :param consumer_secret: Required. The consumer secret for OAuth 1.0 authentication. - :type consumer_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param access_token: Required. The access token for OAuth 1.0 authentication. - :type access_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param access_token_secret: Required. The access token secret for OAuth 1.0 authentication. - :type access_token_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'endpoint': {'required': True}, - 'company_id': {'required': True}, - 'consumer_key': {'required': True}, - 'consumer_secret': {'required': True}, - 'access_token': {'required': True}, - 'access_token_secret': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_properties': {'key': 'typeProperties.connectionProperties', 'type': 'object'}, - 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, - 'company_id': {'key': 'typeProperties.companyId', 'type': 'object'}, - 'consumer_key': {'key': 'typeProperties.consumerKey', 'type': 'object'}, - 'consumer_secret': {'key': 'typeProperties.consumerSecret', 'type': 'SecretBase'}, - 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, - 'access_token_secret': {'key': 'typeProperties.accessTokenSecret', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - endpoint: Any, - company_id: Any, - consumer_key: Any, - consumer_secret: "SecretBase", - access_token: "SecretBase", - access_token_secret: "SecretBase", - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_properties: Optional[Any] = None, - use_encrypted_endpoints: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(QuickBooksLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'QuickBooks' # type: str - self.connection_properties = connection_properties - self.endpoint = endpoint - self.company_id = company_id - self.consumer_key = consumer_key - self.consumer_secret = consumer_secret - self.access_token = access_token - self.access_token_secret = access_token_secret - self.use_encrypted_endpoints = use_encrypted_endpoints - self.encrypted_credential = encrypted_credential - - -class QuickBooksObjectDataset(Dataset): - """QuickBooks server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(QuickBooksObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'QuickBooksObject' # type: str - self.table_name = table_name - - -class QuickBooksSource(TabularSource): - """A copy activity QuickBooks server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(QuickBooksSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'QuickBooksSource' # type: str - self.query = query - - -class RecurrenceSchedule(msrest.serialization.Model): - """The recurrence schedule. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param minutes: The minutes. - :type minutes: list[int] - :param hours: The hours. - :type hours: list[int] - :param week_days: The days of the week. - :type week_days: list[str or ~azure.synapse.artifacts.v2020_12_01.models.DayOfWeek] - :param month_days: The month days. - :type month_days: list[int] - :param monthly_occurrences: The monthly occurrences. - :type monthly_occurrences: - list[~azure.synapse.artifacts.v2020_12_01.models.RecurrenceScheduleOccurrence] - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'minutes': {'key': 'minutes', 'type': '[int]'}, - 'hours': {'key': 'hours', 'type': '[int]'}, - 'week_days': {'key': 'weekDays', 'type': '[str]'}, - 'month_days': {'key': 'monthDays', 'type': '[int]'}, - 'monthly_occurrences': {'key': 'monthlyOccurrences', 'type': '[RecurrenceScheduleOccurrence]'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - minutes: Optional[List[int]] = None, - hours: Optional[List[int]] = None, - week_days: Optional[List[Union[str, "DayOfWeek"]]] = None, - month_days: Optional[List[int]] = None, - monthly_occurrences: Optional[List["RecurrenceScheduleOccurrence"]] = None, - **kwargs - ): - super(RecurrenceSchedule, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.minutes = minutes - self.hours = hours - self.week_days = week_days - self.month_days = month_days - self.monthly_occurrences = monthly_occurrences - - -class RecurrenceScheduleOccurrence(msrest.serialization.Model): - """The recurrence schedule occurrence. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param day: The day of the week. Possible values include: "Sunday", "Monday", "Tuesday", - "Wednesday", "Thursday", "Friday", "Saturday". - :type day: str or ~azure.synapse.artifacts.v2020_12_01.models.DayOfWeek - :param occurrence: The occurrence. - :type occurrence: int - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'day': {'key': 'day', 'type': 'str'}, - 'occurrence': {'key': 'occurrence', 'type': 'int'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - day: Optional[Union[str, "DayOfWeek"]] = None, - occurrence: Optional[int] = None, - **kwargs - ): - super(RecurrenceScheduleOccurrence, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.day = day - self.occurrence = occurrence - - -class RedirectIncompatibleRowSettings(msrest.serialization.Model): - """Redirect incompatible row settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param linked_service_name: Required. Name of the Azure Storage, Storage SAS, or Azure Data - Lake Store linked service used for redirecting incompatible row. Must be specified if - redirectIncompatibleRowSettings is specified. Type: string (or Expression with resultType - string). - :type linked_service_name: any - :param path: The path for storing the redirect incompatible row data. Type: string (or - Expression with resultType string). - :type path: any - """ - - _validation = { - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'object'}, - 'path': {'key': 'path', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: Any, - additional_properties: Optional[Dict[str, Any]] = None, - path: Optional[Any] = None, - **kwargs - ): - super(RedirectIncompatibleRowSettings, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.linked_service_name = linked_service_name - self.path = path - - -class RedshiftUnloadSettings(msrest.serialization.Model): - """The Amazon S3 settings needed for the interim Amazon S3 when copying from Amazon Redshift with unload. With this, data from Amazon Redshift source will be unloaded into S3 first and then copied into the targeted sink from the interim S3. - - All required parameters must be populated in order to send to Azure. - - :param s3_linked_service_name: Required. The name of the Amazon S3 linked service which will be - used for the unload operation when copying from the Amazon Redshift source. - :type s3_linked_service_name: - ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param bucket_name: Required. The bucket of the interim Amazon S3 which will be used to store - the unloaded data from Amazon Redshift source. The bucket must be in the same region as the - Amazon Redshift source. Type: string (or Expression with resultType string). - :type bucket_name: any - """ - - _validation = { - 's3_linked_service_name': {'required': True}, - 'bucket_name': {'required': True}, - } - - _attribute_map = { - 's3_linked_service_name': {'key': 's3LinkedServiceName', 'type': 'LinkedServiceReference'}, - 'bucket_name': {'key': 'bucketName', 'type': 'object'}, - } - - def __init__( - self, - *, - s3_linked_service_name: "LinkedServiceReference", - bucket_name: Any, - **kwargs - ): - super(RedshiftUnloadSettings, self).__init__(**kwargs) - self.s3_linked_service_name = s3_linked_service_name - self.bucket_name = bucket_name - - -class RelationalSource(CopySource): - """A copy activity source for various relational databases. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query: Optional[Any] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(RelationalSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'RelationalSource' # type: str - self.query = query - self.additional_columns = additional_columns - - -class RelationalTableDataset(Dataset): - """The relational table dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The relational table name. Type: string (or Expression with resultType - string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(RelationalTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'RelationalTable' # type: str - self.table_name = table_name - - -class RerunTriggerListResponse(msrest.serialization.Model): - """A list of rerun triggers. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of rerun triggers. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.RerunTriggerResource] - :ivar next_link: The continuation token for getting the next page of results, if any remaining - results exist, null otherwise. - :vartype next_link: str - """ - - _validation = { - 'value': {'required': True}, - 'next_link': {'readonly': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[RerunTriggerResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - *, - value: List["RerunTriggerResource"], - **kwargs - ): - super(RerunTriggerListResponse, self).__init__(**kwargs) - self.value = value - self.next_link = None - - -class RerunTriggerResource(SubResource): - """RerunTrigger resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Properties of the rerun trigger. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.RerunTumblingWindowTrigger - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'RerunTumblingWindowTrigger'}, - } - - def __init__( - self, - *, - properties: "RerunTumblingWindowTrigger", - **kwargs - ): - super(RerunTriggerResource, self).__init__(**kwargs) - self.properties = properties - - -class RerunTumblingWindowTrigger(Trigger): - """Trigger that schedules pipeline reruns for all fixed time interval windows from a requested start time to requested end time. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Trigger type.Constant filled by server. - :type type: str - :param description: Trigger description. - :type description: str - :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are - called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRuntimeState - :param annotations: List of tags that can be used for describing the trigger. - :type annotations: list[any] - :param parent_trigger: Required. The parent trigger reference. - :type parent_trigger: any - :param requested_start_time: Required. The start time for the time period for which restatement - is initiated. Only UTC time is currently supported. - :type requested_start_time: ~datetime.datetime - :param requested_end_time: Required. The end time for the time period for which restatement is - initiated. Only UTC time is currently supported. - :type requested_end_time: ~datetime.datetime - :param rerun_concurrency: Required. The max number of parallel time windows (ready for - execution) for which a rerun is triggered. - :type rerun_concurrency: int - """ - - _validation = { - 'type': {'required': True}, - 'runtime_state': {'readonly': True}, - 'parent_trigger': {'required': True}, - 'requested_start_time': {'required': True}, - 'requested_end_time': {'required': True}, - 'rerun_concurrency': {'required': True, 'maximum': 50, 'minimum': 1}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'parent_trigger': {'key': 'typeProperties.parentTrigger', 'type': 'object'}, - 'requested_start_time': {'key': 'typeProperties.requestedStartTime', 'type': 'iso-8601'}, - 'requested_end_time': {'key': 'typeProperties.requestedEndTime', 'type': 'iso-8601'}, - 'rerun_concurrency': {'key': 'typeProperties.rerunConcurrency', 'type': 'int'}, - } - - def __init__( - self, - *, - parent_trigger: Any, - requested_start_time: datetime.datetime, - requested_end_time: datetime.datetime, - rerun_concurrency: int, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - annotations: Optional[List[Any]] = None, - **kwargs - ): - super(RerunTumblingWindowTrigger, self).__init__(additional_properties=additional_properties, description=description, annotations=annotations, **kwargs) - self.type = 'RerunTumblingWindowTrigger' # type: str - self.parent_trigger = parent_trigger - self.requested_start_time = requested_start_time - self.requested_end_time = requested_end_time - self.rerun_concurrency = rerun_concurrency - - -class RerunTumblingWindowTriggerActionParameters(msrest.serialization.Model): - """Rerun tumbling window trigger Parameters. - - All required parameters must be populated in order to send to Azure. - - :param start_time: Required. The start time for the time period for which restatement is - initiated. Only UTC time is currently supported. - :type start_time: ~datetime.datetime - :param end_time: Required. The end time for the time period for which restatement is initiated. - Only UTC time is currently supported. - :type end_time: ~datetime.datetime - :param max_concurrency: Required. The max number of parallel time windows (ready for execution) - for which a rerun is triggered. - :type max_concurrency: int - """ - - _validation = { - 'start_time': {'required': True}, - 'end_time': {'required': True}, - 'max_concurrency': {'required': True, 'maximum': 50, 'minimum': 1}, - } - - _attribute_map = { - 'start_time': {'key': 'startTime', 'type': 'iso-8601'}, - 'end_time': {'key': 'endTime', 'type': 'iso-8601'}, - 'max_concurrency': {'key': 'maxConcurrency', 'type': 'int'}, - } - - def __init__( - self, - *, - start_time: datetime.datetime, - end_time: datetime.datetime, - max_concurrency: int, - **kwargs - ): - super(RerunTumblingWindowTriggerActionParameters, self).__init__(**kwargs) - self.start_time = start_time - self.end_time = end_time - self.max_concurrency = max_concurrency - - -class ResponsysLinkedService(LinkedService): - """Responsys linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param endpoint: Required. The endpoint of the Responsys server. - :type endpoint: any - :param client_id: Required. The client ID associated with the Responsys application. Type: - string (or Expression with resultType string). - :type client_id: any - :param client_secret: The client secret associated with the Responsys application. Type: string - (or Expression with resultType string). - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. Type: boolean (or Expression with resultType boolean). - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. Type: boolean (or Expression with resultType boolean). - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. Type: boolean (or Expression with resultType - boolean). - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'endpoint': {'required': True}, - 'client_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - endpoint: Any, - client_id: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - client_secret: Optional["SecretBase"] = None, - use_encrypted_endpoints: Optional[Any] = None, - use_host_verification: Optional[Any] = None, - use_peer_verification: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(ResponsysLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Responsys' # type: str - self.endpoint = endpoint - self.client_id = client_id - self.client_secret = client_secret - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential - - -class ResponsysObjectDataset(Dataset): - """Responsys dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(ResponsysObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'ResponsysObject' # type: str - self.table_name = table_name - - -class ResponsysSource(TabularSource): - """A copy activity Responsys source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(ResponsysSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'ResponsysSource' # type: str - self.query = query - - -class RestResourceDataset(Dataset): - """A Rest service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param relative_url: The relative URL to the resource that the RESTful API provides. Type: - string (or Expression with resultType string). - :type relative_url: any - :param request_method: The HTTP method used to call the RESTful API. The default is GET. Type: - string (or Expression with resultType string). - :type request_method: any - :param request_body: The HTTP request body to the RESTful API if requestMethod is POST. Type: - string (or Expression with resultType string). - :type request_body: any - :param additional_headers: The additional HTTP headers in the request to the RESTful API. Type: - string (or Expression with resultType string). - :type additional_headers: any - :param pagination_rules: The pagination rules to compose next page requests. Type: string (or - Expression with resultType string). - :type pagination_rules: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'relative_url': {'key': 'typeProperties.relativeUrl', 'type': 'object'}, - 'request_method': {'key': 'typeProperties.requestMethod', 'type': 'object'}, - 'request_body': {'key': 'typeProperties.requestBody', 'type': 'object'}, - 'additional_headers': {'key': 'typeProperties.additionalHeaders', 'type': 'object'}, - 'pagination_rules': {'key': 'typeProperties.paginationRules', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - relative_url: Optional[Any] = None, - request_method: Optional[Any] = None, - request_body: Optional[Any] = None, - additional_headers: Optional[Any] = None, - pagination_rules: Optional[Any] = None, - **kwargs - ): - super(RestResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'RestResource' # type: str - self.relative_url = relative_url - self.request_method = request_method - self.request_body = request_body - self.additional_headers = additional_headers - self.pagination_rules = pagination_rules - - -class RestServiceLinkedService(LinkedService): - """Rest Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param url: Required. The base URL of the REST service. - :type url: any - :param enable_server_certificate_validation: Whether to validate server side SSL certificate - when connecting to the endpoint.The default value is true. Type: boolean (or Expression with - resultType boolean). - :type enable_server_certificate_validation: any - :param authentication_type: Required. Type of authentication used to connect to the REST - service. Possible values include: "Anonymous", "Basic", "AadServicePrincipal", - "ManagedServiceIdentity". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.RestServiceAuthenticationType - :param user_name: The user name used in Basic authentication type. - :type user_name: any - :param password: The password used in Basic authentication type. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param service_principal_id: The application's client ID used in AadServicePrincipal - authentication type. - :type service_principal_id: any - :param service_principal_key: The application's key used in AadServicePrincipal authentication - type. - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param tenant: The tenant information (domain name or tenant ID) used in AadServicePrincipal - authentication type under which your application resides. - :type tenant: any - :param azure_cloud_type: Indicates the azure cloud type of the service principle auth. Allowed - values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data - factory regions’ cloud type. Type: string (or Expression with resultType string). - :type azure_cloud_type: any - :param aad_resource_id: The resource you are requesting authorization to use. - :type aad_resource_id: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'enable_server_certificate_validation': {'key': 'typeProperties.enableServerCertificateValidation', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'tenant': {'key': 'typeProperties.tenant', 'type': 'object'}, - 'azure_cloud_type': {'key': 'typeProperties.azureCloudType', 'type': 'object'}, - 'aad_resource_id': {'key': 'typeProperties.aadResourceId', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - url: Any, - authentication_type: Union[str, "RestServiceAuthenticationType"], - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - enable_server_certificate_validation: Optional[Any] = None, - user_name: Optional[Any] = None, - password: Optional["SecretBase"] = None, - service_principal_id: Optional[Any] = None, - service_principal_key: Optional["SecretBase"] = None, - tenant: Optional[Any] = None, - azure_cloud_type: Optional[Any] = None, - aad_resource_id: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(RestServiceLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'RestService' # type: str - self.url = url - self.enable_server_certificate_validation = enable_server_certificate_validation - self.authentication_type = authentication_type - self.user_name = user_name - self.password = password - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.tenant = tenant - self.azure_cloud_type = azure_cloud_type - self.aad_resource_id = aad_resource_id - self.encrypted_credential = encrypted_credential - - -class RestSink(CopySink): - """A copy activity Rest service Sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param request_method: The HTTP method used to call the RESTful API. The default is POST. Type: - string (or Expression with resultType string). - :type request_method: any - :param additional_headers: The additional HTTP headers in the request to the RESTful API. Type: - string (or Expression with resultType string). - :type additional_headers: any - :param http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout - to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: any - :param request_interval: The time to await before sending next request, in milliseconds. - :type request_interval: any - :param http_compression_type: Http Compression Type to Send data in compressed format with - Optimal Compression Level, Default is None. And The Only Supported option is Gzip. - :type http_compression_type: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'request_method': {'key': 'requestMethod', 'type': 'object'}, - 'additional_headers': {'key': 'additionalHeaders', 'type': 'object'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - 'request_interval': {'key': 'requestInterval', 'type': 'object'}, - 'http_compression_type': {'key': 'httpCompressionType', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - request_method: Optional[Any] = None, - additional_headers: Optional[Any] = None, - http_request_timeout: Optional[Any] = None, - request_interval: Optional[Any] = None, - http_compression_type: Optional[Any] = None, - **kwargs - ): - super(RestSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'RestSink' # type: str - self.request_method = request_method - self.additional_headers = additional_headers - self.http_request_timeout = http_request_timeout - self.request_interval = request_interval - self.http_compression_type = http_compression_type - - -class RestSource(CopySource): - """A copy activity Rest service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param request_method: The HTTP method used to call the RESTful API. The default is GET. Type: - string (or Expression with resultType string). - :type request_method: any - :param request_body: The HTTP request body to the RESTful API if requestMethod is POST. Type: - string (or Expression with resultType string). - :type request_body: any - :param additional_headers: The additional HTTP headers in the request to the RESTful API. Type: - string (or Expression with resultType string). - :type additional_headers: any - :param pagination_rules: The pagination rules to compose next page requests. Type: string (or - Expression with resultType string). - :type pagination_rules: any - :param http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout - to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: any - :param request_interval: The time to await before sending next page request. - :type request_interval: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'request_method': {'key': 'requestMethod', 'type': 'object'}, - 'request_body': {'key': 'requestBody', 'type': 'object'}, - 'additional_headers': {'key': 'additionalHeaders', 'type': 'object'}, - 'pagination_rules': {'key': 'paginationRules', 'type': 'object'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - 'request_interval': {'key': 'requestInterval', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - request_method: Optional[Any] = None, - request_body: Optional[Any] = None, - additional_headers: Optional[Any] = None, - pagination_rules: Optional[Any] = None, - http_request_timeout: Optional[Any] = None, - request_interval: Optional[Any] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(RestSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'RestSource' # type: str - self.request_method = request_method - self.request_body = request_body - self.additional_headers = additional_headers - self.pagination_rules = pagination_rules - self.http_request_timeout = http_request_timeout - self.request_interval = request_interval - self.additional_columns = additional_columns - - -class RetryPolicy(msrest.serialization.Model): - """Execution policy for an activity. - - :param count: Maximum ordinary retry attempts. Default is 0. Type: integer (or Expression with - resultType integer), minimum: 0. - :type count: any - :param interval_in_seconds: Interval between retries in seconds. Default is 30. - :type interval_in_seconds: int - """ - - _validation = { - 'interval_in_seconds': {'maximum': 86400, 'minimum': 30}, - } - - _attribute_map = { - 'count': {'key': 'count', 'type': 'object'}, - 'interval_in_seconds': {'key': 'intervalInSeconds', 'type': 'int'}, - } - - def __init__( - self, - *, - count: Optional[Any] = None, - interval_in_seconds: Optional[int] = None, - **kwargs - ): - super(RetryPolicy, self).__init__(**kwargs) - self.count = count - self.interval_in_seconds = interval_in_seconds - - -class RunFilterParameters(msrest.serialization.Model): - """Query parameters for listing runs. - - All required parameters must be populated in order to send to Azure. - - :param continuation_token: The continuation token for getting the next page of results. Null - for first page. - :type continuation_token: str - :param last_updated_after: Required. The time at or after which the run event was updated in - 'ISO 8601' format. - :type last_updated_after: ~datetime.datetime - :param last_updated_before: Required. The time at or before which the run event was updated in - 'ISO 8601' format. - :type last_updated_before: ~datetime.datetime - :param filters: List of filters. - :type filters: list[~azure.synapse.artifacts.v2020_12_01.models.RunQueryFilter] - :param order_by: List of OrderBy option. - :type order_by: list[~azure.synapse.artifacts.v2020_12_01.models.RunQueryOrderBy] - """ - - _validation = { - 'last_updated_after': {'required': True}, - 'last_updated_before': {'required': True}, - } - - _attribute_map = { - 'continuation_token': {'key': 'continuationToken', 'type': 'str'}, - 'last_updated_after': {'key': 'lastUpdatedAfter', 'type': 'iso-8601'}, - 'last_updated_before': {'key': 'lastUpdatedBefore', 'type': 'iso-8601'}, - 'filters': {'key': 'filters', 'type': '[RunQueryFilter]'}, - 'order_by': {'key': 'orderBy', 'type': '[RunQueryOrderBy]'}, - } - - def __init__( - self, - *, - last_updated_after: datetime.datetime, - last_updated_before: datetime.datetime, - continuation_token: Optional[str] = None, - filters: Optional[List["RunQueryFilter"]] = None, - order_by: Optional[List["RunQueryOrderBy"]] = None, - **kwargs - ): - super(RunFilterParameters, self).__init__(**kwargs) - self.continuation_token = continuation_token - self.last_updated_after = last_updated_after - self.last_updated_before = last_updated_before - self.filters = filters - self.order_by = order_by - - -class RunQueryFilter(msrest.serialization.Model): - """Query filter option for listing runs. - - All required parameters must be populated in order to send to Azure. - - :param operand: Required. Parameter name to be used for filter. The allowed operands to query - pipeline runs are PipelineName, RunStart, RunEnd and Status; to query activity runs are - ActivityName, ActivityRunStart, ActivityRunEnd, ActivityType and Status, and to query trigger - runs are TriggerName, TriggerRunTimestamp and Status. Possible values include: "PipelineName", - "Status", "RunStart", "RunEnd", "ActivityName", "ActivityRunStart", "ActivityRunEnd", - "ActivityType", "TriggerName", "TriggerRunTimestamp", "RunGroupId", "LatestOnly". - :type operand: str or ~azure.synapse.artifacts.v2020_12_01.models.RunQueryFilterOperand - :param operator: Required. Operator to be used for filter. Possible values include: "Equals", - "NotEquals", "In", "NotIn". - :type operator: str or ~azure.synapse.artifacts.v2020_12_01.models.RunQueryFilterOperator - :param values: Required. List of filter values. - :type values: list[str] - """ - - _validation = { - 'operand': {'required': True}, - 'operator': {'required': True}, - 'values': {'required': True}, - } - - _attribute_map = { - 'operand': {'key': 'operand', 'type': 'str'}, - 'operator': {'key': 'operator', 'type': 'str'}, - 'values': {'key': 'values', 'type': '[str]'}, - } - - def __init__( - self, - *, - operand: Union[str, "RunQueryFilterOperand"], - operator: Union[str, "RunQueryFilterOperator"], - values: List[str], - **kwargs - ): - super(RunQueryFilter, self).__init__(**kwargs) - self.operand = operand - self.operator = operator - self.values = values - - -class RunQueryOrderBy(msrest.serialization.Model): - """An object to provide order by options for listing runs. - - All required parameters must be populated in order to send to Azure. - - :param order_by: Required. Parameter name to be used for order by. The allowed parameters to - order by for pipeline runs are PipelineName, RunStart, RunEnd and Status; for activity runs are - ActivityName, ActivityRunStart, ActivityRunEnd and Status; for trigger runs are TriggerName, - TriggerRunTimestamp and Status. Possible values include: "RunStart", "RunEnd", "PipelineName", - "Status", "ActivityName", "ActivityRunStart", "ActivityRunEnd", "TriggerName", - "TriggerRunTimestamp". - :type order_by: str or ~azure.synapse.artifacts.v2020_12_01.models.RunQueryOrderByField - :param order: Required. Sorting order of the parameter. Possible values include: "ASC", "DESC". - :type order: str or ~azure.synapse.artifacts.v2020_12_01.models.RunQueryOrder - """ - - _validation = { - 'order_by': {'required': True}, - 'order': {'required': True}, - } - - _attribute_map = { - 'order_by': {'key': 'orderBy', 'type': 'str'}, - 'order': {'key': 'order', 'type': 'str'}, - } - - def __init__( - self, - *, - order_by: Union[str, "RunQueryOrderByField"], - order: Union[str, "RunQueryOrder"], - **kwargs - ): - super(RunQueryOrderBy, self).__init__(**kwargs) - self.order_by = order_by - self.order = order - - -class SalesforceLinkedService(LinkedService): - """Linked service for Salesforce. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param environment_url: The URL of Salesforce instance. Default is - 'https://login.salesforce.com'. To copy data from sandbox, specify - 'https://test.salesforce.com'. To copy data from custom domain, specify, for example, - 'https://[domain].my.salesforce.com'. Type: string (or Expression with resultType string). - :type environment_url: any - :param username: The username for Basic authentication of the Salesforce instance. Type: string - (or Expression with resultType string). - :type username: any - :param password: The password for Basic authentication of the Salesforce instance. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param security_token: The security token is optional to remotely access Salesforce instance. - :type security_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param api_version: The Salesforce API version used in ADF. Type: string (or Expression with - resultType string). - :type api_version: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'environment_url': {'key': 'typeProperties.environmentUrl', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'security_token': {'key': 'typeProperties.securityToken', 'type': 'SecretBase'}, - 'api_version': {'key': 'typeProperties.apiVersion', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - environment_url: Optional[Any] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - security_token: Optional["SecretBase"] = None, - api_version: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(SalesforceLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Salesforce' # type: str - self.environment_url = environment_url - self.username = username - self.password = password - self.security_token = security_token - self.api_version = api_version - self.encrypted_credential = encrypted_credential - - -class SalesforceMarketingCloudLinkedService(LinkedService): - """Salesforce Marketing Cloud linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_properties: Properties used to connect to Salesforce Marketing Cloud. It is - mutually exclusive with any other properties in the linked service. Type: object. - :type connection_properties: any - :param client_id: Required. The client ID associated with the Salesforce Marketing Cloud - application. Type: string (or Expression with resultType string). - :type client_id: any - :param client_secret: The client secret associated with the Salesforce Marketing Cloud - application. Type: string (or Expression with resultType string). - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. Type: boolean (or Expression with resultType boolean). - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. Type: boolean (or Expression with resultType boolean). - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. Type: boolean (or Expression with resultType - boolean). - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'client_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_properties': {'key': 'typeProperties.connectionProperties', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - client_id: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_properties: Optional[Any] = None, - client_secret: Optional["SecretBase"] = None, - use_encrypted_endpoints: Optional[Any] = None, - use_host_verification: Optional[Any] = None, - use_peer_verification: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(SalesforceMarketingCloudLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'SalesforceMarketingCloud' # type: str - self.connection_properties = connection_properties - self.client_id = client_id - self.client_secret = client_secret - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential - - -class SalesforceMarketingCloudObjectDataset(Dataset): - """Salesforce Marketing Cloud dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(SalesforceMarketingCloudObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'SalesforceMarketingCloudObject' # type: str - self.table_name = table_name - - -class SalesforceMarketingCloudSource(TabularSource): - """A copy activity Salesforce Marketing Cloud source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(SalesforceMarketingCloudSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'SalesforceMarketingCloudSource' # type: str - self.query = query - - -class SalesforceObjectDataset(Dataset): - """The Salesforce object dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param object_api_name: The Salesforce object API name. Type: string (or Expression with - resultType string). - :type object_api_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'object_api_name': {'key': 'typeProperties.objectApiName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - object_api_name: Optional[Any] = None, - **kwargs - ): - super(SalesforceObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'SalesforceObject' # type: str - self.object_api_name = object_api_name - - -class SalesforceServiceCloudLinkedService(LinkedService): - """Linked service for Salesforce Service Cloud. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param environment_url: The URL of Salesforce Service Cloud instance. Default is - 'https://login.salesforce.com'. To copy data from sandbox, specify - 'https://test.salesforce.com'. To copy data from custom domain, specify, for example, - 'https://[domain].my.salesforce.com'. Type: string (or Expression with resultType string). - :type environment_url: any - :param username: The username for Basic authentication of the Salesforce instance. Type: string - (or Expression with resultType string). - :type username: any - :param password: The password for Basic authentication of the Salesforce instance. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param security_token: The security token is optional to remotely access Salesforce instance. - :type security_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param api_version: The Salesforce API version used in ADF. Type: string (or Expression with - resultType string). - :type api_version: any - :param extended_properties: Extended properties appended to the connection string. Type: string - (or Expression with resultType string). - :type extended_properties: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'environment_url': {'key': 'typeProperties.environmentUrl', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'security_token': {'key': 'typeProperties.securityToken', 'type': 'SecretBase'}, - 'api_version': {'key': 'typeProperties.apiVersion', 'type': 'object'}, - 'extended_properties': {'key': 'typeProperties.extendedProperties', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - environment_url: Optional[Any] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - security_token: Optional["SecretBase"] = None, - api_version: Optional[Any] = None, - extended_properties: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(SalesforceServiceCloudLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'SalesforceServiceCloud' # type: str - self.environment_url = environment_url - self.username = username - self.password = password - self.security_token = security_token - self.api_version = api_version - self.extended_properties = extended_properties - self.encrypted_credential = encrypted_credential - - -class SalesforceServiceCloudObjectDataset(Dataset): - """The Salesforce Service Cloud object dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param object_api_name: The Salesforce Service Cloud object API name. Type: string (or - Expression with resultType string). - :type object_api_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'object_api_name': {'key': 'typeProperties.objectApiName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - object_api_name: Optional[Any] = None, - **kwargs - ): - super(SalesforceServiceCloudObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'SalesforceServiceCloudObject' # type: str - self.object_api_name = object_api_name - - -class SalesforceServiceCloudSink(CopySink): - """A copy activity Salesforce Service Cloud sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param write_behavior: The write behavior for the operation. Default is Insert. Possible values - include: "Insert", "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2020_12_01.models.SalesforceSinkWriteBehavior - :param external_id_field_name: The name of the external ID field for upsert operation. Default - value is 'Id' column. Type: string (or Expression with resultType string). - :type external_id_field_name: any - :param ignore_null_values: The flag indicating whether or not to ignore null values from input - dataset (except key fields) during write operation. Default value is false. If set it to true, - it means ADF will leave the data in the destination object unchanged when doing upsert/update - operation and insert defined default value when doing insert operation, versus ADF will update - the data in the destination object to NULL when doing upsert/update operation and insert NULL - value when doing insert operation. Type: boolean (or Expression with resultType boolean). - :type ignore_null_values: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, - 'external_id_field_name': {'key': 'externalIdFieldName', 'type': 'object'}, - 'ignore_null_values': {'key': 'ignoreNullValues', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - write_behavior: Optional[Union[str, "SalesforceSinkWriteBehavior"]] = None, - external_id_field_name: Optional[Any] = None, - ignore_null_values: Optional[Any] = None, - **kwargs - ): - super(SalesforceServiceCloudSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'SalesforceServiceCloudSink' # type: str - self.write_behavior = write_behavior - self.external_id_field_name = external_id_field_name - self.ignore_null_values = ignore_null_values - - -class SalesforceServiceCloudSource(CopySource): - """A copy activity Salesforce Service Cloud source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - :param read_behavior: The read behavior for the operation. Default is Query. Possible values - include: "Query", "QueryAll". - :type read_behavior: str or - ~azure.synapse.artifacts.v2020_12_01.models.SalesforceSourceReadBehavior - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'read_behavior': {'key': 'readBehavior', 'type': 'str'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query: Optional[Any] = None, - read_behavior: Optional[Union[str, "SalesforceSourceReadBehavior"]] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(SalesforceServiceCloudSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'SalesforceServiceCloudSource' # type: str - self.query = query - self.read_behavior = read_behavior - self.additional_columns = additional_columns - - -class SalesforceSink(CopySink): - """A copy activity Salesforce sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param write_behavior: The write behavior for the operation. Default is Insert. Possible values - include: "Insert", "Upsert". - :type write_behavior: str or - ~azure.synapse.artifacts.v2020_12_01.models.SalesforceSinkWriteBehavior - :param external_id_field_name: The name of the external ID field for upsert operation. Default - value is 'Id' column. Type: string (or Expression with resultType string). - :type external_id_field_name: any - :param ignore_null_values: The flag indicating whether or not to ignore null values from input - dataset (except key fields) during write operation. Default value is false. If set it to true, - it means ADF will leave the data in the destination object unchanged when doing upsert/update - operation and insert defined default value when doing insert operation, versus ADF will update - the data in the destination object to NULL when doing upsert/update operation and insert NULL - value when doing insert operation. Type: boolean (or Expression with resultType boolean). - :type ignore_null_values: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, - 'external_id_field_name': {'key': 'externalIdFieldName', 'type': 'object'}, - 'ignore_null_values': {'key': 'ignoreNullValues', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - write_behavior: Optional[Union[str, "SalesforceSinkWriteBehavior"]] = None, - external_id_field_name: Optional[Any] = None, - ignore_null_values: Optional[Any] = None, - **kwargs - ): - super(SalesforceSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'SalesforceSink' # type: str - self.write_behavior = write_behavior - self.external_id_field_name = external_id_field_name - self.ignore_null_values = ignore_null_values - - -class SalesforceSource(TabularSource): - """A copy activity Salesforce source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - :param read_behavior: The read behavior for the operation. Default is Query. Possible values - include: "Query", "QueryAll". - :type read_behavior: str or - ~azure.synapse.artifacts.v2020_12_01.models.SalesforceSourceReadBehavior - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'read_behavior': {'key': 'readBehavior', 'type': 'str'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - read_behavior: Optional[Union[str, "SalesforceSourceReadBehavior"]] = None, - **kwargs - ): - super(SalesforceSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'SalesforceSource' # type: str - self.query = query - self.read_behavior = read_behavior - - -class SapBwCubeDataset(Dataset): - """The SAP BW cube dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - **kwargs - ): - super(SapBwCubeDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'SapBwCube' # type: str - - -class SapBWLinkedService(LinkedService): - """SAP Business Warehouse Linked Service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param server: Required. Host name of the SAP BW instance. Type: string (or Expression with - resultType string). - :type server: any - :param system_number: Required. System number of the BW system. (Usually a two-digit decimal - number represented as a string.) Type: string (or Expression with resultType string). - :type system_number: any - :param client_id: Required. Client ID of the client on the BW system. (Usually a three-digit - decimal number represented as a string) Type: string (or Expression with resultType string). - :type client_id: any - :param user_name: Username to access the SAP BW server. Type: string (or Expression with - resultType string). - :type user_name: any - :param password: Password to access the SAP BW server. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'server': {'required': True}, - 'system_number': {'required': True}, - 'client_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'system_number': {'key': 'typeProperties.systemNumber', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - server: Any, - system_number: Any, - client_id: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - user_name: Optional[Any] = None, - password: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(SapBWLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'SapBW' # type: str - self.server = server - self.system_number = system_number - self.client_id = client_id - self.user_name = user_name - self.password = password - self.encrypted_credential = encrypted_credential - - -class SapBwSource(TabularSource): - """A copy activity source for SapBW server via MDX. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: MDX query. Type: string (or Expression with resultType string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(SapBwSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'SapBwSource' # type: str - self.query = query - - -class SapCloudForCustomerLinkedService(LinkedService): - """Linked service for SAP Cloud for Customer. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param url: Required. The URL of SAP Cloud for Customer OData API. For example, - '[https://[tenantname].crm.ondemand.com/sap/c4c/odata/v1]'. Type: string (or Expression with - resultType string). - :type url: any - :param username: The username for Basic authentication. Type: string (or Expression with - resultType string). - :type username: any - :param password: The password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Either encryptedCredential or - username/password must be provided. Type: string (or Expression with resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - url: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(SapCloudForCustomerLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'SapCloudForCustomer' # type: str - self.url = url - self.username = username - self.password = password - self.encrypted_credential = encrypted_credential - - -class SapCloudForCustomerResourceDataset(Dataset): - """The path of the SAP Cloud for Customer OData entity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param path: Required. The path of the SAP Cloud for Customer OData entity. Type: string (or - Expression with resultType string). - :type path: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'path': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'path': {'key': 'typeProperties.path', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - path: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - **kwargs - ): - super(SapCloudForCustomerResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'SapCloudForCustomerResource' # type: str - self.path = path - - -class SapCloudForCustomerSink(CopySink): - """A copy activity SAP Cloud for Customer sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param write_behavior: The write behavior for the operation. Default is 'Insert'. Possible - values include: "Insert", "Update". - :type write_behavior: str or - ~azure.synapse.artifacts.v2020_12_01.models.SapCloudForCustomerSinkWriteBehavior - :param http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout - to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - write_behavior: Optional[Union[str, "SapCloudForCustomerSinkWriteBehavior"]] = None, - http_request_timeout: Optional[Any] = None, - **kwargs - ): - super(SapCloudForCustomerSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'SapCloudForCustomerSink' # type: str - self.write_behavior = write_behavior - self.http_request_timeout = http_request_timeout - - -class SapCloudForCustomerSource(TabularSource): - """A copy activity source for SAP Cloud for Customer source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: SAP Cloud for Customer OData query. For example, "$top=1". Type: string (or - Expression with resultType string). - :type query: any - :param http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout - to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - http_request_timeout: Optional[Any] = None, - **kwargs - ): - super(SapCloudForCustomerSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'SapCloudForCustomerSource' # type: str - self.query = query - self.http_request_timeout = http_request_timeout - - -class SapEccLinkedService(LinkedService): - """Linked service for SAP ERP Central Component(SAP ECC). - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param url: Required. The URL of SAP ECC OData API. For example, - '[https://hostname:port/sap/opu/odata/sap/servicename/]'. Type: string (or Expression with - resultType string). - :type url: str - :param username: The username for Basic authentication. Type: string (or Expression with - resultType string). - :type username: str - :param password: The password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Either encryptedCredential or - username/password must be provided. Type: string (or Expression with resultType string). - :type encrypted_credential: str - """ - - _validation = { - 'type': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'url': {'key': 'typeProperties.url', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'str'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'str'}, - } - - def __init__( - self, - *, - url: str, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - username: Optional[str] = None, - password: Optional["SecretBase"] = None, - encrypted_credential: Optional[str] = None, - **kwargs - ): - super(SapEccLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'SapEcc' # type: str - self.url = url - self.username = username - self.password = password - self.encrypted_credential = encrypted_credential - - -class SapEccResourceDataset(Dataset): - """The path of the SAP ECC OData entity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param path: Required. The path of the SAP ECC OData entity. Type: string (or Expression with - resultType string). - :type path: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'path': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'path': {'key': 'typeProperties.path', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - path: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - **kwargs - ): - super(SapEccResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'SapEccResource' # type: str - self.path = path - - -class SapEccSource(TabularSource): - """A copy activity source for SAP ECC source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: SAP ECC OData query. For example, "$top=1". Type: string (or Expression with - resultType string). - :type query: any - :param http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout - to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - http_request_timeout: Optional[Any] = None, - **kwargs - ): - super(SapEccSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'SapEccSource' # type: str - self.query = query - self.http_request_timeout = http_request_timeout - - -class SapHanaLinkedService(LinkedService): - """SAP HANA Linked Service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: SAP HANA ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param server: Required. Host name of the SAP HANA server. Type: string (or Expression with - resultType string). - :type server: any - :param authentication_type: The authentication type to be used to connect to the SAP HANA - server. Possible values include: "Basic", "Windows". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.SapHanaAuthenticationType - :param user_name: Username to access the SAP HANA server. Type: string (or Expression with - resultType string). - :type user_name: any - :param password: Password to access the SAP HANA server. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'server': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - server: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_string: Optional[Any] = None, - authentication_type: Optional[Union[str, "SapHanaAuthenticationType"]] = None, - user_name: Optional[Any] = None, - password: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(SapHanaLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'SapHana' # type: str - self.connection_string = connection_string - self.server = server - self.authentication_type = authentication_type - self.user_name = user_name - self.password = password - self.encrypted_credential = encrypted_credential - - -class SapHanaPartitionSettings(msrest.serialization.Model): - """The settings that will be leveraged for SAP HANA source partitioning. - - :param partition_column_name: The name of the column that will be used for proceeding range - partitioning. Type: string (or Expression with resultType string). - :type partition_column_name: any - """ - - _attribute_map = { - 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, - } - - def __init__( - self, - *, - partition_column_name: Optional[Any] = None, - **kwargs - ): - super(SapHanaPartitionSettings, self).__init__(**kwargs) - self.partition_column_name = partition_column_name - - -class SapHanaSource(TabularSource): - """A copy activity source for SAP HANA source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: SAP HANA Sql query. Type: string (or Expression with resultType string). - :type query: any - :param packet_size: The packet size of data read from SAP HANA. Type: integer(or Expression - with resultType integer). - :type packet_size: any - :param partition_option: The partition mechanism that will be used for SAP HANA read in - parallel. Possible values include: "None", "PhysicalPartitionsOfTable", "SapHanaDynamicRange". - :type partition_option: str or - ~azure.synapse.artifacts.v2020_12_01.models.SapHanaPartitionOption - :param partition_settings: The settings that will be leveraged for SAP HANA source - partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.SapHanaPartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'packet_size': {'key': 'packetSize', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'SapHanaPartitionSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - packet_size: Optional[Any] = None, - partition_option: Optional[Union[str, "SapHanaPartitionOption"]] = None, - partition_settings: Optional["SapHanaPartitionSettings"] = None, - **kwargs - ): - super(SapHanaSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'SapHanaSource' # type: str - self.query = query - self.packet_size = packet_size - self.partition_option = partition_option - self.partition_settings = partition_settings - - -class SapHanaTableDataset(Dataset): - """SAP HANA Table properties. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param schema_type_properties_schema: The schema name of SAP HANA. Type: string (or Expression - with resultType string). - :type schema_type_properties_schema: any - :param table: The table name of SAP HANA. Type: string (or Expression with resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - schema_type_properties_schema: Optional[Any] = None, - table: Optional[Any] = None, - **kwargs - ): - super(SapHanaTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'SapHanaTable' # type: str - self.schema_type_properties_schema = schema_type_properties_schema - self.table = table - - -class SapOpenHubLinkedService(LinkedService): - """SAP Business Warehouse Open Hub Destination Linked Service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param server: Required. Host name of the SAP BW instance where the open hub destination is - located. Type: string (or Expression with resultType string). - :type server: any - :param system_number: Required. System number of the BW system where the open hub destination - is located. (Usually a two-digit decimal number represented as a string.) Type: string (or - Expression with resultType string). - :type system_number: any - :param client_id: Required. Client ID of the client on the BW system where the open hub - destination is located. (Usually a three-digit decimal number represented as a string) Type: - string (or Expression with resultType string). - :type client_id: any - :param language: Language of the BW system where the open hub destination is located. The - default value is EN. Type: string (or Expression with resultType string). - :type language: any - :param system_id: SystemID of the SAP system where the table is located. Type: string (or - Expression with resultType string). - :type system_id: any - :param user_name: Username to access the SAP BW server where the open hub destination is - located. Type: string (or Expression with resultType string). - :type user_name: any - :param password: Password to access the SAP BW server where the open hub destination is - located. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param message_server: The hostname of the SAP Message Server. Type: string (or Expression with - resultType string). - :type message_server: any - :param message_server_service: The service name or port number of the Message Server. Type: - string (or Expression with resultType string). - :type message_server_service: any - :param logon_group: The Logon Group for the SAP System. Type: string (or Expression with - resultType string). - :type logon_group: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'server': {'required': True}, - 'system_number': {'required': True}, - 'client_id': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'system_number': {'key': 'typeProperties.systemNumber', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'language': {'key': 'typeProperties.language', 'type': 'object'}, - 'system_id': {'key': 'typeProperties.systemId', 'type': 'object'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'message_server': {'key': 'typeProperties.messageServer', 'type': 'object'}, - 'message_server_service': {'key': 'typeProperties.messageServerService', 'type': 'object'}, - 'logon_group': {'key': 'typeProperties.logonGroup', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - server: Any, - system_number: Any, - client_id: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - language: Optional[Any] = None, - system_id: Optional[Any] = None, - user_name: Optional[Any] = None, - password: Optional["SecretBase"] = None, - message_server: Optional[Any] = None, - message_server_service: Optional[Any] = None, - logon_group: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(SapOpenHubLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'SapOpenHub' # type: str - self.server = server - self.system_number = system_number - self.client_id = client_id - self.language = language - self.system_id = system_id - self.user_name = user_name - self.password = password - self.message_server = message_server - self.message_server_service = message_server_service - self.logon_group = logon_group - self.encrypted_credential = encrypted_credential - - -class SapOpenHubSource(TabularSource): - """A copy activity source for SAP Business Warehouse Open Hub Destination source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param exclude_last_request: Whether to exclude the records of the last request. The default - value is true. Type: boolean (or Expression with resultType boolean). - :type exclude_last_request: any - :param base_request_id: The ID of request for delta loading. Once it is set, only data with - requestId larger than the value of this property will be retrieved. The default value is 0. - Type: integer (or Expression with resultType integer ). - :type base_request_id: any - :param custom_rfc_read_table_function_module: Specifies the custom RFC function module that - will be used to read data from SAP Table. Type: string (or Expression with resultType string). - :type custom_rfc_read_table_function_module: any - :param sap_data_column_delimiter: The single character that will be used as delimiter passed to - SAP RFC as well as splitting the output data retrieved. Type: string (or Expression with - resultType string). - :type sap_data_column_delimiter: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'exclude_last_request': {'key': 'excludeLastRequest', 'type': 'object'}, - 'base_request_id': {'key': 'baseRequestId', 'type': 'object'}, - 'custom_rfc_read_table_function_module': {'key': 'customRfcReadTableFunctionModule', 'type': 'object'}, - 'sap_data_column_delimiter': {'key': 'sapDataColumnDelimiter', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - exclude_last_request: Optional[Any] = None, - base_request_id: Optional[Any] = None, - custom_rfc_read_table_function_module: Optional[Any] = None, - sap_data_column_delimiter: Optional[Any] = None, - **kwargs - ): - super(SapOpenHubSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'SapOpenHubSource' # type: str - self.exclude_last_request = exclude_last_request - self.base_request_id = base_request_id - self.custom_rfc_read_table_function_module = custom_rfc_read_table_function_module - self.sap_data_column_delimiter = sap_data_column_delimiter - - -class SapOpenHubTableDataset(Dataset): - """Sap Business Warehouse Open Hub Destination Table properties. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param open_hub_destination_name: Required. The name of the Open Hub Destination with - destination type as Database Table. Type: string (or Expression with resultType string). - :type open_hub_destination_name: any - :param exclude_last_request: Whether to exclude the records of the last request. The default - value is true. Type: boolean (or Expression with resultType boolean). - :type exclude_last_request: any - :param base_request_id: The ID of request for delta loading. Once it is set, only data with - requestId larger than the value of this property will be retrieved. The default value is 0. - Type: integer (or Expression with resultType integer ). - :type base_request_id: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'open_hub_destination_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'open_hub_destination_name': {'key': 'typeProperties.openHubDestinationName', 'type': 'object'}, - 'exclude_last_request': {'key': 'typeProperties.excludeLastRequest', 'type': 'object'}, - 'base_request_id': {'key': 'typeProperties.baseRequestId', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - open_hub_destination_name: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - exclude_last_request: Optional[Any] = None, - base_request_id: Optional[Any] = None, - **kwargs - ): - super(SapOpenHubTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'SapOpenHubTable' # type: str - self.open_hub_destination_name = open_hub_destination_name - self.exclude_last_request = exclude_last_request - self.base_request_id = base_request_id - - -class SapTableLinkedService(LinkedService): - """SAP Table Linked Service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param server: Host name of the SAP instance where the table is located. Type: string (or - Expression with resultType string). - :type server: any - :param system_number: System number of the SAP system where the table is located. (Usually a - two-digit decimal number represented as a string.) Type: string (or Expression with resultType - string). - :type system_number: any - :param client_id: Client ID of the client on the SAP system where the table is located. - (Usually a three-digit decimal number represented as a string) Type: string (or Expression with - resultType string). - :type client_id: any - :param language: Language of the SAP system where the table is located. The default value is - EN. Type: string (or Expression with resultType string). - :type language: any - :param system_id: SystemID of the SAP system where the table is located. Type: string (or - Expression with resultType string). - :type system_id: any - :param user_name: Username to access the SAP server where the table is located. Type: string - (or Expression with resultType string). - :type user_name: any - :param password: Password to access the SAP server where the table is located. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param message_server: The hostname of the SAP Message Server. Type: string (or Expression with - resultType string). - :type message_server: any - :param message_server_service: The service name or port number of the Message Server. Type: - string (or Expression with resultType string). - :type message_server_service: any - :param snc_mode: SNC activation indicator to access the SAP server where the table is located. - Must be either 0 (off) or 1 (on). Type: string (or Expression with resultType string). - :type snc_mode: any - :param snc_my_name: Initiator's SNC name to access the SAP server where the table is located. - Type: string (or Expression with resultType string). - :type snc_my_name: any - :param snc_partner_name: Communication partner's SNC name to access the SAP server where the - table is located. Type: string (or Expression with resultType string). - :type snc_partner_name: any - :param snc_library_path: External security product's library to access the SAP server where the - table is located. Type: string (or Expression with resultType string). - :type snc_library_path: any - :param snc_qop: SNC Quality of Protection. Allowed value include: 1, 2, 3, 8, 9. Type: string - (or Expression with resultType string). - :type snc_qop: any - :param logon_group: The Logon Group for the SAP System. Type: string (or Expression with - resultType string). - :type logon_group: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'system_number': {'key': 'typeProperties.systemNumber', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'language': {'key': 'typeProperties.language', 'type': 'object'}, - 'system_id': {'key': 'typeProperties.systemId', 'type': 'object'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'message_server': {'key': 'typeProperties.messageServer', 'type': 'object'}, - 'message_server_service': {'key': 'typeProperties.messageServerService', 'type': 'object'}, - 'snc_mode': {'key': 'typeProperties.sncMode', 'type': 'object'}, - 'snc_my_name': {'key': 'typeProperties.sncMyName', 'type': 'object'}, - 'snc_partner_name': {'key': 'typeProperties.sncPartnerName', 'type': 'object'}, - 'snc_library_path': {'key': 'typeProperties.sncLibraryPath', 'type': 'object'}, - 'snc_qop': {'key': 'typeProperties.sncQop', 'type': 'object'}, - 'logon_group': {'key': 'typeProperties.logonGroup', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - server: Optional[Any] = None, - system_number: Optional[Any] = None, - client_id: Optional[Any] = None, - language: Optional[Any] = None, - system_id: Optional[Any] = None, - user_name: Optional[Any] = None, - password: Optional["SecretBase"] = None, - message_server: Optional[Any] = None, - message_server_service: Optional[Any] = None, - snc_mode: Optional[Any] = None, - snc_my_name: Optional[Any] = None, - snc_partner_name: Optional[Any] = None, - snc_library_path: Optional[Any] = None, - snc_qop: Optional[Any] = None, - logon_group: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(SapTableLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'SapTable' # type: str - self.server = server - self.system_number = system_number - self.client_id = client_id - self.language = language - self.system_id = system_id - self.user_name = user_name - self.password = password - self.message_server = message_server - self.message_server_service = message_server_service - self.snc_mode = snc_mode - self.snc_my_name = snc_my_name - self.snc_partner_name = snc_partner_name - self.snc_library_path = snc_library_path - self.snc_qop = snc_qop - self.logon_group = logon_group - self.encrypted_credential = encrypted_credential - - -class SapTablePartitionSettings(msrest.serialization.Model): - """The settings that will be leveraged for SAP table source partitioning. - - :param partition_column_name: The name of the column that will be used for proceeding range - partitioning. Type: string (or Expression with resultType string). - :type partition_column_name: any - :param partition_upper_bound: The maximum value of column specified in partitionColumnName that - will be used for proceeding range partitioning. Type: string (or Expression with resultType - string). - :type partition_upper_bound: any - :param partition_lower_bound: The minimum value of column specified in partitionColumnName that - will be used for proceeding range partitioning. Type: string (or Expression with resultType - string). - :type partition_lower_bound: any - :param max_partitions_number: The maximum value of partitions the table will be split into. - Type: integer (or Expression with resultType string). - :type max_partitions_number: any - """ - - _attribute_map = { - 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, - 'partition_upper_bound': {'key': 'partitionUpperBound', 'type': 'object'}, - 'partition_lower_bound': {'key': 'partitionLowerBound', 'type': 'object'}, - 'max_partitions_number': {'key': 'maxPartitionsNumber', 'type': 'object'}, - } - - def __init__( - self, - *, - partition_column_name: Optional[Any] = None, - partition_upper_bound: Optional[Any] = None, - partition_lower_bound: Optional[Any] = None, - max_partitions_number: Optional[Any] = None, - **kwargs - ): - super(SapTablePartitionSettings, self).__init__(**kwargs) - self.partition_column_name = partition_column_name - self.partition_upper_bound = partition_upper_bound - self.partition_lower_bound = partition_lower_bound - self.max_partitions_number = max_partitions_number - - -class SapTableResourceDataset(Dataset): - """SAP Table Resource properties. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: Required. The name of the SAP Table. Type: string (or Expression with - resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'table_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - table_name: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - **kwargs - ): - super(SapTableResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'SapTableResource' # type: str - self.table_name = table_name - - -class SapTableSource(TabularSource): - """A copy activity source for SAP Table source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param row_count: The number of rows to be retrieved. Type: integer(or Expression with - resultType integer). - :type row_count: any - :param row_skips: The number of rows that will be skipped. Type: integer (or Expression with - resultType integer). - :type row_skips: any - :param rfc_table_fields: The fields of the SAP table that will be retrieved. For example, - column0, column1. Type: string (or Expression with resultType string). - :type rfc_table_fields: any - :param rfc_table_options: The options for the filtering of the SAP Table. For example, COLUMN0 - EQ SOME VALUE. Type: string (or Expression with resultType string). - :type rfc_table_options: any - :param batch_size: Specifies the maximum number of rows that will be retrieved at a time when - retrieving data from SAP Table. Type: integer (or Expression with resultType integer). - :type batch_size: any - :param custom_rfc_read_table_function_module: Specifies the custom RFC function module that - will be used to read data from SAP Table. Type: string (or Expression with resultType string). - :type custom_rfc_read_table_function_module: any - :param sap_data_column_delimiter: The single character that will be used as delimiter passed to - SAP RFC as well as splitting the output data retrieved. Type: string (or Expression with - resultType string). - :type sap_data_column_delimiter: any - :param partition_option: The partition mechanism that will be used for SAP table read in - parallel. Possible values include: "None", "PartitionOnInt", "PartitionOnCalendarYear", - "PartitionOnCalendarMonth", "PartitionOnCalendarDate", "PartitionOnTime". - :type partition_option: str or - ~azure.synapse.artifacts.v2020_12_01.models.SapTablePartitionOption - :param partition_settings: The settings that will be leveraged for SAP table source - partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.SapTablePartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'row_count': {'key': 'rowCount', 'type': 'object'}, - 'row_skips': {'key': 'rowSkips', 'type': 'object'}, - 'rfc_table_fields': {'key': 'rfcTableFields', 'type': 'object'}, - 'rfc_table_options': {'key': 'rfcTableOptions', 'type': 'object'}, - 'batch_size': {'key': 'batchSize', 'type': 'object'}, - 'custom_rfc_read_table_function_module': {'key': 'customRfcReadTableFunctionModule', 'type': 'object'}, - 'sap_data_column_delimiter': {'key': 'sapDataColumnDelimiter', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'SapTablePartitionSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - row_count: Optional[Any] = None, - row_skips: Optional[Any] = None, - rfc_table_fields: Optional[Any] = None, - rfc_table_options: Optional[Any] = None, - batch_size: Optional[Any] = None, - custom_rfc_read_table_function_module: Optional[Any] = None, - sap_data_column_delimiter: Optional[Any] = None, - partition_option: Optional[Union[str, "SapTablePartitionOption"]] = None, - partition_settings: Optional["SapTablePartitionSettings"] = None, - **kwargs - ): - super(SapTableSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'SapTableSource' # type: str - self.row_count = row_count - self.row_skips = row_skips - self.rfc_table_fields = rfc_table_fields - self.rfc_table_options = rfc_table_options - self.batch_size = batch_size - self.custom_rfc_read_table_function_module = custom_rfc_read_table_function_module - self.sap_data_column_delimiter = sap_data_column_delimiter - self.partition_option = partition_option - self.partition_settings = partition_settings - - -class ScheduleTrigger(MultiplePipelineTrigger): - """Trigger that creates pipeline runs periodically, on schedule. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Trigger type.Constant filled by server. - :type type: str - :param description: Trigger description. - :type description: str - :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are - called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRuntimeState - :param annotations: List of tags that can be used for describing the trigger. - :type annotations: list[any] - :param pipelines: Pipelines that need to be started. - :type pipelines: list[~azure.synapse.artifacts.v2020_12_01.models.TriggerPipelineReference] - :param recurrence: Required. Recurrence schedule configuration. - :type recurrence: ~azure.synapse.artifacts.v2020_12_01.models.ScheduleTriggerRecurrence - """ - - _validation = { - 'type': {'required': True}, - 'runtime_state': {'readonly': True}, - 'recurrence': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, - 'recurrence': {'key': 'typeProperties.recurrence', 'type': 'ScheduleTriggerRecurrence'}, - } - - def __init__( - self, - *, - recurrence: "ScheduleTriggerRecurrence", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - annotations: Optional[List[Any]] = None, - pipelines: Optional[List["TriggerPipelineReference"]] = None, - **kwargs - ): - super(ScheduleTrigger, self).__init__(additional_properties=additional_properties, description=description, annotations=annotations, pipelines=pipelines, **kwargs) - self.type = 'ScheduleTrigger' # type: str - self.recurrence = recurrence - - -class ScheduleTriggerRecurrence(msrest.serialization.Model): - """The workflow trigger recurrence. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param frequency: The frequency. Possible values include: "NotSpecified", "Minute", "Hour", - "Day", "Week", "Month", "Year". - :type frequency: str or ~azure.synapse.artifacts.v2020_12_01.models.RecurrenceFrequency - :param interval: The interval. - :type interval: int - :param start_time: The start time. - :type start_time: ~datetime.datetime - :param end_time: The end time. - :type end_time: ~datetime.datetime - :param time_zone: The time zone. - :type time_zone: str - :param schedule: The recurrence schedule. - :type schedule: ~azure.synapse.artifacts.v2020_12_01.models.RecurrenceSchedule - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'frequency': {'key': 'frequency', 'type': 'str'}, - 'interval': {'key': 'interval', 'type': 'int'}, - 'start_time': {'key': 'startTime', 'type': 'iso-8601'}, - 'end_time': {'key': 'endTime', 'type': 'iso-8601'}, - 'time_zone': {'key': 'timeZone', 'type': 'str'}, - 'schedule': {'key': 'schedule', 'type': 'RecurrenceSchedule'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - frequency: Optional[Union[str, "RecurrenceFrequency"]] = None, - interval: Optional[int] = None, - start_time: Optional[datetime.datetime] = None, - end_time: Optional[datetime.datetime] = None, - time_zone: Optional[str] = None, - schedule: Optional["RecurrenceSchedule"] = None, - **kwargs - ): - super(ScheduleTriggerRecurrence, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.frequency = frequency - self.interval = interval - self.start_time = start_time - self.end_time = end_time - self.time_zone = time_zone - self.schedule = schedule - - -class ScriptAction(msrest.serialization.Model): - """Custom script action to run on HDI ondemand cluster once it's up. - - All required parameters must be populated in order to send to Azure. - - :param name: Required. The user provided name of the script action. - :type name: str - :param uri: Required. The URI for the script action. - :type uri: str - :param roles: Required. The node types on which the script action should be executed. Possible - values include: "Headnode", "Workernode", "Zookeeper". - :type roles: str or ~azure.synapse.artifacts.v2020_12_01.models.HdiNodeTypes - :param parameters: The parameters for the script action. - :type parameters: str - """ - - _validation = { - 'name': {'required': True}, - 'uri': {'required': True}, - 'roles': {'required': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'uri': {'key': 'uri', 'type': 'str'}, - 'roles': {'key': 'roles', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': 'str'}, - } - - def __init__( - self, - *, - name: str, - uri: str, - roles: Union[str, "HdiNodeTypes"], - parameters: Optional[str] = None, - **kwargs - ): - super(ScriptAction, self).__init__(**kwargs) - self.name = name - self.uri = uri - self.roles = roles - self.parameters = parameters - - -class SecureString(SecretBase): - """Azure Synapse secure string definition. The string value will be masked with asterisks '*' during Get or List API calls. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Type of the secret.Constant filled by server. - :type type: str - :param value: Required. Value of secure string. - :type value: str - """ - - _validation = { - 'type': {'required': True}, - 'value': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'value': {'key': 'value', 'type': 'str'}, - } - - def __init__( - self, - *, - value: str, - **kwargs - ): - super(SecureString, self).__init__(**kwargs) - self.type = 'SecureString' # type: str - self.value = value - - -class SelfDependencyTumblingWindowTriggerReference(DependencyReference): - """Self referenced tumbling window trigger dependency. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. The type of dependency reference.Constant filled by server. - :type type: str - :param offset: Required. Timespan applied to the start time of a tumbling window when - evaluating dependency. - :type offset: str - :param size: The size of the window when evaluating the dependency. If undefined the frequency - of the tumbling window will be used. - :type size: str - """ - - _validation = { - 'type': {'required': True}, - 'offset': {'required': True, 'max_length': 15, 'min_length': 8, 'pattern': r'-((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))'}, - 'size': {'max_length': 15, 'min_length': 8, 'pattern': r'((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))'}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'offset': {'key': 'offset', 'type': 'str'}, - 'size': {'key': 'size', 'type': 'str'}, - } - - def __init__( - self, - *, - offset: str, - size: Optional[str] = None, - **kwargs - ): - super(SelfDependencyTumblingWindowTriggerReference, self).__init__(**kwargs) - self.type = 'SelfDependencyTumblingWindowTriggerReference' # type: str - self.offset = offset - self.size = size - - -class SelfHostedIntegrationRuntime(IntegrationRuntime): - """Self-hosted integration runtime. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of integration runtime.Constant filled by server. Possible values - include: "Managed", "SelfHosted". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeType - :param description: Integration runtime description. - :type description: str - :param linked_info: Linked integration runtime type from data factory. - :type linked_info: ~azure.synapse.artifacts.v2020_12_01.models.LinkedIntegrationRuntimeType - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'linked_info': {'key': 'typeProperties.linkedInfo', 'type': 'LinkedIntegrationRuntimeType'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - linked_info: Optional["LinkedIntegrationRuntimeType"] = None, - **kwargs - ): - super(SelfHostedIntegrationRuntime, self).__init__(additional_properties=additional_properties, description=description, **kwargs) - self.type = 'SelfHosted' # type: str - self.linked_info = linked_info - - -class ServiceNowLinkedService(LinkedService): - """ServiceNow server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param endpoint: Required. The endpoint of the ServiceNow server. (i.e. - :code:``.service-now.com). - :type endpoint: any - :param authentication_type: Required. The authentication type to use. Possible values include: - "Basic", "OAuth2". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.ServiceNowAuthenticationType - :param username: The user name used to connect to the ServiceNow server for Basic and OAuth2 - authentication. - :type username: any - :param password: The password corresponding to the user name for Basic and OAuth2 - authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param client_id: The client id for OAuth2 authentication. - :type client_id: any - :param client_secret: The client secret for OAuth2 authentication. - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'endpoint': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - endpoint: Any, - authentication_type: Union[str, "ServiceNowAuthenticationType"], - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - client_id: Optional[Any] = None, - client_secret: Optional["SecretBase"] = None, - use_encrypted_endpoints: Optional[Any] = None, - use_host_verification: Optional[Any] = None, - use_peer_verification: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(ServiceNowLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'ServiceNow' # type: str - self.endpoint = endpoint - self.authentication_type = authentication_type - self.username = username - self.password = password - self.client_id = client_id - self.client_secret = client_secret - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential - - -class ServiceNowObjectDataset(Dataset): - """ServiceNow server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(ServiceNowObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'ServiceNowObject' # type: str - self.table_name = table_name - - -class ServiceNowSource(TabularSource): - """A copy activity ServiceNow server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(ServiceNowSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'ServiceNowSource' # type: str - self.query = query - - -class SetVariableActivity(ControlActivity): - """Set value for a Variable. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param variable_name: Name of the variable whose value needs to be set. - :type variable_name: str - :param value: Value to be set. Could be a static value or Expression. - :type value: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'variable_name': {'key': 'typeProperties.variableName', 'type': 'str'}, - 'value': {'key': 'typeProperties.value', 'type': 'object'}, - } - - def __init__( - self, - *, - name: str, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - variable_name: Optional[str] = None, - value: Optional[Any] = None, - **kwargs - ): - super(SetVariableActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) - self.type = 'SetVariable' # type: str - self.variable_name = variable_name - self.value = value - - -class SftpLocation(DatasetLocation): - """The location of SFTP dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage location.Constant filled by server. - :type type: str - :param folder_path: Specify the folder path of dataset. Type: string (or Expression with - resultType string). - :type folder_path: any - :param file_name: Specify the file name of dataset. Type: string (or Expression with resultType - string). - :type file_name: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'folder_path': {'key': 'folderPath', 'type': 'object'}, - 'file_name': {'key': 'fileName', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - folder_path: Optional[Any] = None, - file_name: Optional[Any] = None, - **kwargs - ): - super(SftpLocation, self).__init__(additional_properties=additional_properties, folder_path=folder_path, file_name=file_name, **kwargs) - self.type = 'SftpLocation' # type: str - - -class SftpReadSettings(StoreReadSettings): - """Sftp read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param recursive: If true, files under the folder path will be read recursively. Default is - true. Type: boolean (or Expression with resultType boolean). - :type recursive: any - :param wildcard_folder_path: Sftp wildcardFolderPath. Type: string (or Expression with - resultType string). - :type wildcard_folder_path: any - :param wildcard_file_name: Sftp wildcardFileName. Type: string (or Expression with resultType - string). - :type wildcard_file_name: any - :param enable_partition_discovery: Indicates whether to enable partition discovery. - :type enable_partition_discovery: bool - :param partition_root_path: Specify the root path where partition discovery starts from. Type: - string (or Expression with resultType string). - :type partition_root_path: any - :param file_list_path: Point to a text file that lists each file (relative path to the path - configured in the dataset) that you want to copy. Type: string (or Expression with resultType - string). - :type file_list_path: any - :param delete_files_after_completion: Indicates whether the source files need to be deleted - after copy completion. Default is false. Type: boolean (or Expression with resultType boolean). - :type delete_files_after_completion: any - :param modified_datetime_start: The start of file's modified datetime. Type: string (or - Expression with resultType string). - :type modified_datetime_start: any - :param modified_datetime_end: The end of file's modified datetime. Type: string (or Expression - with resultType string). - :type modified_datetime_end: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'recursive': {'key': 'recursive', 'type': 'object'}, - 'wildcard_folder_path': {'key': 'wildcardFolderPath', 'type': 'object'}, - 'wildcard_file_name': {'key': 'wildcardFileName', 'type': 'object'}, - 'enable_partition_discovery': {'key': 'enablePartitionDiscovery', 'type': 'bool'}, - 'partition_root_path': {'key': 'partitionRootPath', 'type': 'object'}, - 'file_list_path': {'key': 'fileListPath', 'type': 'object'}, - 'delete_files_after_completion': {'key': 'deleteFilesAfterCompletion', 'type': 'object'}, - 'modified_datetime_start': {'key': 'modifiedDatetimeStart', 'type': 'object'}, - 'modified_datetime_end': {'key': 'modifiedDatetimeEnd', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - recursive: Optional[Any] = None, - wildcard_folder_path: Optional[Any] = None, - wildcard_file_name: Optional[Any] = None, - enable_partition_discovery: Optional[bool] = None, - partition_root_path: Optional[Any] = None, - file_list_path: Optional[Any] = None, - delete_files_after_completion: Optional[Any] = None, - modified_datetime_start: Optional[Any] = None, - modified_datetime_end: Optional[Any] = None, - **kwargs - ): - super(SftpReadSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'SftpReadSettings' # type: str - self.recursive = recursive - self.wildcard_folder_path = wildcard_folder_path - self.wildcard_file_name = wildcard_file_name - self.enable_partition_discovery = enable_partition_discovery - self.partition_root_path = partition_root_path - self.file_list_path = file_list_path - self.delete_files_after_completion = delete_files_after_completion - self.modified_datetime_start = modified_datetime_start - self.modified_datetime_end = modified_datetime_end - - -class SftpServerLinkedService(LinkedService): - """A linked service for an SSH File Transfer Protocol (SFTP) server. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The SFTP server host name. Type: string (or Expression with resultType - string). - :type host: any - :param port: The TCP port number that the SFTP server uses to listen for client connections. - Default value is 22. Type: integer (or Expression with resultType integer), minimum: 0. - :type port: any - :param authentication_type: The authentication type to be used to connect to the FTP server. - Possible values include: "Basic", "SshPublicKey". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.SftpAuthenticationType - :param user_name: The username used to log on to the SFTP server. Type: string (or Expression - with resultType string). - :type user_name: any - :param password: Password to logon the SFTP server for Basic authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - :param private_key_path: The SSH private key file path for SshPublicKey authentication. Only - valid for on-premises copy. For on-premises copy with SshPublicKey authentication, either - PrivateKeyPath or PrivateKeyContent should be specified. SSH private key should be OpenSSH - format. Type: string (or Expression with resultType string). - :type private_key_path: any - :param private_key_content: Base64 encoded SSH private key content for SshPublicKey - authentication. For on-premises copy with SshPublicKey authentication, either PrivateKeyPath or - PrivateKeyContent should be specified. SSH private key should be OpenSSH format. - :type private_key_content: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param pass_phrase: The password to decrypt the SSH private key if the SSH private key is - encrypted. - :type pass_phrase: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param skip_host_key_validation: If true, skip the SSH host key validation. Default value is - false. Type: boolean (or Expression with resultType boolean). - :type skip_host_key_validation: any - :param host_key_fingerprint: The host key finger-print of the SFTP server. When - SkipHostKeyValidation is false, HostKeyFingerprint should be specified. Type: string (or - Expression with resultType string). - :type host_key_fingerprint: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - 'private_key_path': {'key': 'typeProperties.privateKeyPath', 'type': 'object'}, - 'private_key_content': {'key': 'typeProperties.privateKeyContent', 'type': 'SecretBase'}, - 'pass_phrase': {'key': 'typeProperties.passPhrase', 'type': 'SecretBase'}, - 'skip_host_key_validation': {'key': 'typeProperties.skipHostKeyValidation', 'type': 'object'}, - 'host_key_fingerprint': {'key': 'typeProperties.hostKeyFingerprint', 'type': 'object'}, - } - - def __init__( - self, - *, - host: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - port: Optional[Any] = None, - authentication_type: Optional[Union[str, "SftpAuthenticationType"]] = None, - user_name: Optional[Any] = None, - password: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - private_key_path: Optional[Any] = None, - private_key_content: Optional["SecretBase"] = None, - pass_phrase: Optional["SecretBase"] = None, - skip_host_key_validation: Optional[Any] = None, - host_key_fingerprint: Optional[Any] = None, - **kwargs - ): - super(SftpServerLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Sftp' # type: str - self.host = host - self.port = port - self.authentication_type = authentication_type - self.user_name = user_name - self.password = password - self.encrypted_credential = encrypted_credential - self.private_key_path = private_key_path - self.private_key_content = private_key_content - self.pass_phrase = pass_phrase - self.skip_host_key_validation = skip_host_key_validation - self.host_key_fingerprint = host_key_fingerprint - - -class SftpWriteSettings(StoreWriteSettings): - """Sftp write settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The write setting type.Constant filled by server. - :type type: str - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param copy_behavior: The type of copy behavior for copy sink. - :type copy_behavior: any - :param operation_timeout: Specifies the timeout for writing each chunk to SFTP server. Default - value: 01:00:00 (one hour). Type: string (or Expression with resultType string). - :type operation_timeout: any - :param use_temp_file_rename: Upload to temporary file(s) and rename. Disable this option if - your SFTP server doesn't support rename operation. Type: boolean (or Expression with resultType - boolean). - :type use_temp_file_rename: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, - 'operation_timeout': {'key': 'operationTimeout', 'type': 'object'}, - 'use_temp_file_rename': {'key': 'useTempFileRename', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - max_concurrent_connections: Optional[Any] = None, - copy_behavior: Optional[Any] = None, - operation_timeout: Optional[Any] = None, - use_temp_file_rename: Optional[Any] = None, - **kwargs - ): - super(SftpWriteSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) - self.type = 'SftpWriteSettings' # type: str - self.operation_timeout = operation_timeout - self.use_temp_file_rename = use_temp_file_rename - - -class SharePointOnlineListLinkedService(LinkedService): - """SharePoint Online List linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param site_url: Required. The URL of the SharePoint Online site. For example, - https://contoso.sharepoint.com/sites/siteName. Type: string (or Expression with resultType - string). - :type site_url: any - :param tenant_id: Required. The tenant ID under which your application resides. You can find it - from Azure portal Active Directory overview page. Type: string (or Expression with resultType - string). - :type tenant_id: any - :param service_principal_id: Required. The application (client) ID of your application - registered in Azure Active Directory. Make sure to grant SharePoint site permission to this - application. Type: string (or Expression with resultType string). - :type service_principal_id: any - :param service_principal_key: Required. The client secret of your application registered in - Azure Active Directory. Type: string (or Expression with resultType string). - :type service_principal_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'site_url': {'required': True}, - 'tenant_id': {'required': True}, - 'service_principal_id': {'required': True}, - 'service_principal_key': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'site_url': {'key': 'typeProperties.siteUrl', 'type': 'object'}, - 'tenant_id': {'key': 'typeProperties.tenantId', 'type': 'object'}, - 'service_principal_id': {'key': 'typeProperties.servicePrincipalId', 'type': 'object'}, - 'service_principal_key': {'key': 'typeProperties.servicePrincipalKey', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - site_url: Any, - tenant_id: Any, - service_principal_id: Any, - service_principal_key: "SecretBase", - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(SharePointOnlineListLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'SharePointOnlineList' # type: str - self.site_url = site_url - self.tenant_id = tenant_id - self.service_principal_id = service_principal_id - self.service_principal_key = service_principal_key - self.encrypted_credential = encrypted_credential - - -class SharePointOnlineListResourceDataset(Dataset): - """The sharepoint online list resource dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param list_name: The name of the SharePoint Online list. Type: string (or Expression with - resultType string). - :type list_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'list_name': {'key': 'typeProperties.listName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - list_name: Optional[Any] = None, - **kwargs - ): - super(SharePointOnlineListResourceDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'SharePointOnlineListResource' # type: str - self.list_name = list_name - - -class SharePointOnlineListSource(CopySource): - """A copy activity source for sharePoint online list source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: The OData query to filter the data in SharePoint Online list. For example, - "$top=1". Type: string (or Expression with resultType string). - :type query: any - :param http_request_timeout: The wait time to get a response from SharePoint Online. Default - value is 5 minutes (00:05:00). Type: string (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type http_request_timeout: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query: Optional[Any] = None, - http_request_timeout: Optional[Any] = None, - **kwargs - ): - super(SharePointOnlineListSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'SharePointOnlineListSource' # type: str - self.query = query - self.http_request_timeout = http_request_timeout - - -class ShopifyLinkedService(LinkedService): - """Shopify Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. The endpoint of the Shopify server. (i.e. mystore.myshopify.com). - :type host: any - :param access_token: The API access token that can be used to access Shopify’s data. The token - won't expire if it is offline mode. - :type access_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - host: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - access_token: Optional["SecretBase"] = None, - use_encrypted_endpoints: Optional[Any] = None, - use_host_verification: Optional[Any] = None, - use_peer_verification: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(ShopifyLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Shopify' # type: str - self.host = host - self.access_token = access_token - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential - - -class ShopifyObjectDataset(Dataset): - """Shopify Service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(ShopifyObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'ShopifyObject' # type: str - self.table_name = table_name - - -class ShopifySource(TabularSource): - """A copy activity Shopify Service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(ShopifySource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'ShopifySource' # type: str - self.query = query - - -class SkipErrorFile(msrest.serialization.Model): - """Skip error file. - - :param file_missing: Skip if file is deleted by other client during copy. Default is true. - Type: boolean (or Expression with resultType boolean). - :type file_missing: any - :param data_inconsistency: Skip if source/sink file changed by other concurrent write. Default - is false. Type: boolean (or Expression with resultType boolean). - :type data_inconsistency: any - """ - - _attribute_map = { - 'file_missing': {'key': 'fileMissing', 'type': 'object'}, - 'data_inconsistency': {'key': 'dataInconsistency', 'type': 'object'}, - } - - def __init__( - self, - *, - file_missing: Optional[Any] = None, - data_inconsistency: Optional[Any] = None, - **kwargs - ): - super(SkipErrorFile, self).__init__(**kwargs) - self.file_missing = file_missing - self.data_inconsistency = data_inconsistency - - -class Sku(msrest.serialization.Model): - """SQL pool SKU. - - :param tier: The service tier. - :type tier: str - :param name: The SKU name. - :type name: str - :param capacity: If the SKU supports scale out/in then the capacity integer should be included. - If scale out/in is not possible for the resource this may be omitted. - :type capacity: int - """ - - _attribute_map = { - 'tier': {'key': 'tier', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'capacity': {'key': 'capacity', 'type': 'int'}, - } - - def __init__( - self, - *, - tier: Optional[str] = None, - name: Optional[str] = None, - capacity: Optional[int] = None, - **kwargs - ): - super(Sku, self).__init__(**kwargs) - self.tier = tier - self.name = name - self.capacity = capacity - - -class SnowflakeDataset(Dataset): - """The snowflake dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param schema_type_properties_schema: The schema name of the Snowflake database. Type: string - (or Expression with resultType string). - :type schema_type_properties_schema: any - :param table: The table name of the Snowflake database. Type: string (or Expression with - resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - schema_type_properties_schema: Optional[Any] = None, - table: Optional[Any] = None, - **kwargs - ): - super(SnowflakeDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'SnowflakeTable' # type: str - self.schema_type_properties_schema = schema_type_properties_schema - self.table = table - - -class SnowflakeExportCopyCommand(ExportSettings): - """Snowflake export command settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The export setting type.Constant filled by server. - :type type: str - :param additional_copy_options: Additional copy options directly passed to snowflake Copy - Command. Type: key value pairs (value should be string type) (or Expression with resultType - object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": - "'HH24:MI:SS.FF'" }. - :type additional_copy_options: dict[str, any] - :param additional_format_options: Additional format options directly passed to snowflake Copy - Command. Type: key value pairs (value should be string type) (or Expression with resultType - object). Example: "additionalFormatOptions": { "OVERWRITE": "TRUE", "MAX_FILE_SIZE": "'FALSE'" - }. - :type additional_format_options: dict[str, any] - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'additional_copy_options': {'key': 'additionalCopyOptions', 'type': '{object}'}, - 'additional_format_options': {'key': 'additionalFormatOptions', 'type': '{object}'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - additional_copy_options: Optional[Dict[str, Any]] = None, - additional_format_options: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(SnowflakeExportCopyCommand, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'SnowflakeExportCopyCommand' # type: str - self.additional_copy_options = additional_copy_options - self.additional_format_options = additional_format_options - - -class SnowflakeImportCopyCommand(ImportSettings): - """Snowflake import command settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The import setting type.Constant filled by server. - :type type: str - :param additional_copy_options: Additional copy options directly passed to snowflake Copy - Command. Type: key value pairs (value should be string type) (or Expression with resultType - object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": - "'HH24:MI:SS.FF'" }. - :type additional_copy_options: dict[str, any] - :param additional_format_options: Additional format options directly passed to snowflake Copy - Command. Type: key value pairs (value should be string type) (or Expression with resultType - object). Example: "additionalFormatOptions": { "FORCE": "TRUE", "LOAD_UNCERTAIN_FILES": - "'FALSE'" }. - :type additional_format_options: dict[str, any] - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'additional_copy_options': {'key': 'additionalCopyOptions', 'type': '{object}'}, - 'additional_format_options': {'key': 'additionalFormatOptions', 'type': '{object}'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - additional_copy_options: Optional[Dict[str, Any]] = None, - additional_format_options: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(SnowflakeImportCopyCommand, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'SnowflakeImportCopyCommand' # type: str - self.additional_copy_options = additional_copy_options - self.additional_format_options = additional_format_options - - -class SnowflakeLinkedService(LinkedService): - """Snowflake linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The connection string of snowflake. Type: string, - SecureString. - :type connection_string: any - :param password: The Azure key vault secret reference of password in connection string. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - connection_string: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - password: Optional["AzureKeyVaultSecretReference"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(SnowflakeLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Snowflake' # type: str - self.connection_string = connection_string - self.password = password - self.encrypted_credential = encrypted_credential - - -class SnowflakeSink(CopySink): - """A copy activity snowflake sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType - string). - :type pre_copy_script: any - :param import_settings: Snowflake import settings. - :type import_settings: ~azure.synapse.artifacts.v2020_12_01.models.SnowflakeImportCopyCommand - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - 'import_settings': {'key': 'importSettings', 'type': 'SnowflakeImportCopyCommand'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - pre_copy_script: Optional[Any] = None, - import_settings: Optional["SnowflakeImportCopyCommand"] = None, - **kwargs - ): - super(SnowflakeSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'SnowflakeSink' # type: str - self.pre_copy_script = pre_copy_script - self.import_settings = import_settings - - -class SnowflakeSource(CopySource): - """A copy activity snowflake source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query: Snowflake Sql query. Type: string (or Expression with resultType string). - :type query: any - :param export_settings: Snowflake export settings. - :type export_settings: ~azure.synapse.artifacts.v2020_12_01.models.SnowflakeExportCopyCommand - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'export_settings': {'key': 'exportSettings', 'type': 'SnowflakeExportCopyCommand'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query: Optional[Any] = None, - export_settings: Optional["SnowflakeExportCopyCommand"] = None, - **kwargs - ): - super(SnowflakeSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'SnowflakeSource' # type: str - self.query = query - self.export_settings = export_settings - - -class SparkBatchJob(msrest.serialization.Model): - """SparkBatchJob. - - All required parameters must be populated in order to send to Azure. - - :param livy_info: - :type livy_info: ~azure.synapse.artifacts.v2020_12_01.models.SparkBatchJobState - :param name: The batch name. - :type name: str - :param workspace_name: The workspace name. - :type workspace_name: str - :param spark_pool_name: The Spark pool name. - :type spark_pool_name: str - :param submitter_name: The submitter name. - :type submitter_name: str - :param submitter_id: The submitter identifier. - :type submitter_id: str - :param artifact_id: The artifact identifier. - :type artifact_id: str - :param job_type: The job type. Possible values include: "SparkBatch", "SparkSession". - :type job_type: str or ~azure.synapse.artifacts.v2020_12_01.models.SparkJobType - :param result: The Spark batch job result. Possible values include: "Uncertain", "Succeeded", - "Failed", "Cancelled". - :type result: str or ~azure.synapse.artifacts.v2020_12_01.models.SparkBatchJobResultType - :param scheduler: The scheduler information. - :type scheduler: ~azure.synapse.artifacts.v2020_12_01.models.SparkScheduler - :param plugin: The plugin information. - :type plugin: ~azure.synapse.artifacts.v2020_12_01.models.SparkServicePlugin - :param errors: The error information. - :type errors: list[~azure.synapse.artifacts.v2020_12_01.models.SparkServiceError] - :param tags: A set of tags. The tags. - :type tags: dict[str, str] - :param id: Required. The session Id. - :type id: int - :param app_id: The application id of this session. - :type app_id: str - :param app_info: The detailed application info. - :type app_info: dict[str, str] - :param state: The batch state. - :type state: str - :param log_lines: The log lines. - :type log_lines: list[str] - """ - - _validation = { - 'id': {'required': True}, - } - - _attribute_map = { - 'livy_info': {'key': 'livyInfo', 'type': 'SparkBatchJobState'}, - 'name': {'key': 'name', 'type': 'str'}, - 'workspace_name': {'key': 'workspaceName', 'type': 'str'}, - 'spark_pool_name': {'key': 'sparkPoolName', 'type': 'str'}, - 'submitter_name': {'key': 'submitterName', 'type': 'str'}, - 'submitter_id': {'key': 'submitterId', 'type': 'str'}, - 'artifact_id': {'key': 'artifactId', 'type': 'str'}, - 'job_type': {'key': 'jobType', 'type': 'str'}, - 'result': {'key': 'result', 'type': 'str'}, - 'scheduler': {'key': 'schedulerInfo', 'type': 'SparkScheduler'}, - 'plugin': {'key': 'pluginInfo', 'type': 'SparkServicePlugin'}, - 'errors': {'key': 'errorInfo', 'type': '[SparkServiceError]'}, - 'tags': {'key': 'tags', 'type': '{str}'}, - 'id': {'key': 'id', 'type': 'int'}, - 'app_id': {'key': 'appId', 'type': 'str'}, - 'app_info': {'key': 'appInfo', 'type': '{str}'}, - 'state': {'key': 'state', 'type': 'str'}, - 'log_lines': {'key': 'log', 'type': '[str]'}, - } - - def __init__( - self, - *, - id: int, - livy_info: Optional["SparkBatchJobState"] = None, - name: Optional[str] = None, - workspace_name: Optional[str] = None, - spark_pool_name: Optional[str] = None, - submitter_name: Optional[str] = None, - submitter_id: Optional[str] = None, - artifact_id: Optional[str] = None, - job_type: Optional[Union[str, "SparkJobType"]] = None, - result: Optional[Union[str, "SparkBatchJobResultType"]] = None, - scheduler: Optional["SparkScheduler"] = None, - plugin: Optional["SparkServicePlugin"] = None, - errors: Optional[List["SparkServiceError"]] = None, - tags: Optional[Dict[str, str]] = None, - app_id: Optional[str] = None, - app_info: Optional[Dict[str, str]] = None, - state: Optional[str] = None, - log_lines: Optional[List[str]] = None, - **kwargs - ): - super(SparkBatchJob, self).__init__(**kwargs) - self.livy_info = livy_info - self.name = name - self.workspace_name = workspace_name - self.spark_pool_name = spark_pool_name - self.submitter_name = submitter_name - self.submitter_id = submitter_id - self.artifact_id = artifact_id - self.job_type = job_type - self.result = result - self.scheduler = scheduler - self.plugin = plugin - self.errors = errors - self.tags = tags - self.id = id - self.app_id = app_id - self.app_info = app_info - self.state = state - self.log_lines = log_lines - - -class SparkBatchJobState(msrest.serialization.Model): - """SparkBatchJobState. - - :param not_started_at: the time that at which "not_started" livy state was first seen. - :type not_started_at: ~datetime.datetime - :param starting_at: the time that at which "starting" livy state was first seen. - :type starting_at: ~datetime.datetime - :param running_at: the time that at which "running" livy state was first seen. - :type running_at: ~datetime.datetime - :param dead_at: time that at which "dead" livy state was first seen. - :type dead_at: ~datetime.datetime - :param success_at: the time that at which "success" livy state was first seen. - :type success_at: ~datetime.datetime - :param terminated_at: the time that at which "killed" livy state was first seen. - :type terminated_at: ~datetime.datetime - :param recovering_at: the time that at which "recovering" livy state was first seen. - :type recovering_at: ~datetime.datetime - :param current_state: the Spark job state. - :type current_state: str - :param job_creation_request: - :type job_creation_request: ~azure.synapse.artifacts.v2020_12_01.models.SparkRequest - """ - - _attribute_map = { - 'not_started_at': {'key': 'notStartedAt', 'type': 'iso-8601'}, - 'starting_at': {'key': 'startingAt', 'type': 'iso-8601'}, - 'running_at': {'key': 'runningAt', 'type': 'iso-8601'}, - 'dead_at': {'key': 'deadAt', 'type': 'iso-8601'}, - 'success_at': {'key': 'successAt', 'type': 'iso-8601'}, - 'terminated_at': {'key': 'killedAt', 'type': 'iso-8601'}, - 'recovering_at': {'key': 'recoveringAt', 'type': 'iso-8601'}, - 'current_state': {'key': 'currentState', 'type': 'str'}, - 'job_creation_request': {'key': 'jobCreationRequest', 'type': 'SparkRequest'}, - } - - def __init__( - self, - *, - not_started_at: Optional[datetime.datetime] = None, - starting_at: Optional[datetime.datetime] = None, - running_at: Optional[datetime.datetime] = None, - dead_at: Optional[datetime.datetime] = None, - success_at: Optional[datetime.datetime] = None, - terminated_at: Optional[datetime.datetime] = None, - recovering_at: Optional[datetime.datetime] = None, - current_state: Optional[str] = None, - job_creation_request: Optional["SparkRequest"] = None, - **kwargs - ): - super(SparkBatchJobState, self).__init__(**kwargs) - self.not_started_at = not_started_at - self.starting_at = starting_at - self.running_at = running_at - self.dead_at = dead_at - self.success_at = success_at - self.terminated_at = terminated_at - self.recovering_at = recovering_at - self.current_state = current_state - self.job_creation_request = job_creation_request - - -class SparkJobDefinition(msrest.serialization.Model): - """Spark job definition. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param description: The description of the Spark job definition. - :type description: str - :param target_big_data_pool: Required. Big data pool reference. - :type target_big_data_pool: ~azure.synapse.artifacts.v2020_12_01.models.BigDataPoolReference - :param required_spark_version: The required Spark version of the application. - :type required_spark_version: str - :param language: The language of the Spark application. - :type language: str - :param job_properties: Required. The properties of the Spark job. - :type job_properties: ~azure.synapse.artifacts.v2020_12_01.models.SparkJobProperties - """ - - _validation = { - 'target_big_data_pool': {'required': True}, - 'job_properties': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'description', 'type': 'str'}, - 'target_big_data_pool': {'key': 'targetBigDataPool', 'type': 'BigDataPoolReference'}, - 'required_spark_version': {'key': 'requiredSparkVersion', 'type': 'str'}, - 'language': {'key': 'language', 'type': 'str'}, - 'job_properties': {'key': 'jobProperties', 'type': 'SparkJobProperties'}, - } - - def __init__( - self, - *, - target_big_data_pool: "BigDataPoolReference", - job_properties: "SparkJobProperties", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - required_spark_version: Optional[str] = None, - language: Optional[str] = None, - **kwargs - ): - super(SparkJobDefinition, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.description = description - self.target_big_data_pool = target_big_data_pool - self.required_spark_version = required_spark_version - self.language = language - self.job_properties = job_properties - - -class SparkJobDefinitionResource(SubResource): - """Spark job definition resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Properties of spark job definition. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.SparkJobDefinition - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'SparkJobDefinition'}, - } - - def __init__( - self, - *, - properties: "SparkJobDefinition", - **kwargs - ): - super(SparkJobDefinitionResource, self).__init__(**kwargs) - self.properties = properties - - -class SparkJobDefinitionsListResponse(msrest.serialization.Model): - """A list of spark job definitions resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of spark job definitions. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.SparkJobDefinitionResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[SparkJobDefinitionResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - *, - value: List["SparkJobDefinitionResource"], - next_link: Optional[str] = None, - **kwargs - ): - super(SparkJobDefinitionsListResponse, self).__init__(**kwargs) - self.value = value - self.next_link = next_link - - -class SparkJobProperties(msrest.serialization.Model): - """The properties of the Spark job. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: The name of the job. - :type name: str - :param file: Required. File containing the application to execute. - :type file: str - :param class_name: Main class for Java/Scala application. - :type class_name: str - :param conf: Spark configuration properties. - :type conf: any - :param args: Command line arguments for the application. - :type args: list[str] - :param jars: Jars to be used in this job. - :type jars: list[str] - :param files: files to be used in this job. - :type files: list[str] - :param archives: Archives to be used in this job. - :type archives: list[str] - :param driver_memory: Required. Amount of memory to use for the driver process. - :type driver_memory: str - :param driver_cores: Required. Number of cores to use for the driver. - :type driver_cores: int - :param executor_memory: Required. Amount of memory to use per executor process. - :type executor_memory: str - :param executor_cores: Required. Number of cores to use for each executor. - :type executor_cores: int - :param num_executors: Required. Number of executors to launch for this job. - :type num_executors: int - """ - - _validation = { - 'file': {'required': True}, - 'driver_memory': {'required': True}, - 'driver_cores': {'required': True}, - 'executor_memory': {'required': True}, - 'executor_cores': {'required': True}, - 'num_executors': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'file': {'key': 'file', 'type': 'str'}, - 'class_name': {'key': 'className', 'type': 'str'}, - 'conf': {'key': 'conf', 'type': 'object'}, - 'args': {'key': 'args', 'type': '[str]'}, - 'jars': {'key': 'jars', 'type': '[str]'}, - 'files': {'key': 'files', 'type': '[str]'}, - 'archives': {'key': 'archives', 'type': '[str]'}, - 'driver_memory': {'key': 'driverMemory', 'type': 'str'}, - 'driver_cores': {'key': 'driverCores', 'type': 'int'}, - 'executor_memory': {'key': 'executorMemory', 'type': 'str'}, - 'executor_cores': {'key': 'executorCores', 'type': 'int'}, - 'num_executors': {'key': 'numExecutors', 'type': 'int'}, - } - - def __init__( - self, - *, - file: str, - driver_memory: str, - driver_cores: int, - executor_memory: str, - executor_cores: int, - num_executors: int, - additional_properties: Optional[Dict[str, Any]] = None, - name: Optional[str] = None, - class_name: Optional[str] = None, - conf: Optional[Any] = None, - args: Optional[List[str]] = None, - jars: Optional[List[str]] = None, - files: Optional[List[str]] = None, - archives: Optional[List[str]] = None, - **kwargs - ): - super(SparkJobProperties, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.name = name - self.file = file - self.class_name = class_name - self.conf = conf - self.args = args - self.jars = jars - self.files = files - self.archives = archives - self.driver_memory = driver_memory - self.driver_cores = driver_cores - self.executor_memory = executor_memory - self.executor_cores = executor_cores - self.num_executors = num_executors - - -class SparkLinkedService(LinkedService): - """Spark Server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param host: Required. IP address or host name of the Spark server. - :type host: any - :param port: Required. The TCP port that the Spark server uses to listen for client - connections. - :type port: any - :param server_type: The type of Spark server. Possible values include: "SharkServer", - "SharkServer2", "SparkThriftServer". - :type server_type: str or ~azure.synapse.artifacts.v2020_12_01.models.SparkServerType - :param thrift_transport_protocol: The transport protocol to use in the Thrift layer. Possible - values include: "Binary", "SASL", "HTTP ". - :type thrift_transport_protocol: str or - ~azure.synapse.artifacts.v2020_12_01.models.SparkThriftTransportProtocol - :param authentication_type: Required. The authentication method used to access the Spark - server. Possible values include: "Anonymous", "Username", "UsernameAndPassword", - "WindowsAzureHDInsightService". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.SparkAuthenticationType - :param username: The user name that you use to access Spark Server. - :type username: any - :param password: The password corresponding to the user name that you provided in the Username - field. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param http_path: The partial URL corresponding to the Spark server. - :type http_path: any - :param enable_ssl: Specifies whether the connections to the server are encrypted using SSL. The - default value is false. - :type enable_ssl: any - :param trusted_cert_path: The full path of the .pem file containing trusted CA certificates for - verifying the server when connecting over SSL. This property can only be set when using SSL on - self-hosted IR. The default value is the cacerts.pem file installed with the IR. - :type trusted_cert_path: any - :param use_system_trust_store: Specifies whether to use a CA certificate from the system trust - store or from a specified PEM file. The default value is false. - :type use_system_trust_store: any - :param allow_host_name_cn_mismatch: Specifies whether to require a CA-issued SSL certificate - name to match the host name of the server when connecting over SSL. The default value is false. - :type allow_host_name_cn_mismatch: any - :param allow_self_signed_server_cert: Specifies whether to allow self-signed certificates from - the server. The default value is false. - :type allow_self_signed_server_cert: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'port': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'port': {'key': 'typeProperties.port', 'type': 'object'}, - 'server_type': {'key': 'typeProperties.serverType', 'type': 'str'}, - 'thrift_transport_protocol': {'key': 'typeProperties.thriftTransportProtocol', 'type': 'str'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'http_path': {'key': 'typeProperties.httpPath', 'type': 'object'}, - 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, - 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, - 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, - 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, - 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - host: Any, - port: Any, - authentication_type: Union[str, "SparkAuthenticationType"], - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - server_type: Optional[Union[str, "SparkServerType"]] = None, - thrift_transport_protocol: Optional[Union[str, "SparkThriftTransportProtocol"]] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - http_path: Optional[Any] = None, - enable_ssl: Optional[Any] = None, - trusted_cert_path: Optional[Any] = None, - use_system_trust_store: Optional[Any] = None, - allow_host_name_cn_mismatch: Optional[Any] = None, - allow_self_signed_server_cert: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(SparkLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Spark' # type: str - self.host = host - self.port = port - self.server_type = server_type - self.thrift_transport_protocol = thrift_transport_protocol - self.authentication_type = authentication_type - self.username = username - self.password = password - self.http_path = http_path - self.enable_ssl = enable_ssl - self.trusted_cert_path = trusted_cert_path - self.use_system_trust_store = use_system_trust_store - self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch - self.allow_self_signed_server_cert = allow_self_signed_server_cert - self.encrypted_credential = encrypted_credential - - -class SparkObjectDataset(Dataset): - """Spark Server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The table name of the Spark. Type: string (or Expression with resultType string). - :type table: any - :param schema_type_properties_schema: The schema name of the Spark. Type: string (or Expression - with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - table: Optional[Any] = None, - schema_type_properties_schema: Optional[Any] = None, - **kwargs - ): - super(SparkObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'SparkObject' # type: str - self.table_name = table_name - self.table = table - self.schema_type_properties_schema = schema_type_properties_schema - - -class SparkRequest(msrest.serialization.Model): - """SparkRequest. - - :param name: - :type name: str - :param file: - :type file: str - :param class_name: - :type class_name: str - :param arguments: - :type arguments: list[str] - :param jars: - :type jars: list[str] - :param python_files: - :type python_files: list[str] - :param files: - :type files: list[str] - :param archives: - :type archives: list[str] - :param configuration: Dictionary of :code:``. - :type configuration: dict[str, str] - :param driver_memory: - :type driver_memory: str - :param driver_cores: - :type driver_cores: int - :param executor_memory: - :type executor_memory: str - :param executor_cores: - :type executor_cores: int - :param executor_count: - :type executor_count: int - """ - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'file': {'key': 'file', 'type': 'str'}, - 'class_name': {'key': 'className', 'type': 'str'}, - 'arguments': {'key': 'args', 'type': '[str]'}, - 'jars': {'key': 'jars', 'type': '[str]'}, - 'python_files': {'key': 'pyFiles', 'type': '[str]'}, - 'files': {'key': 'files', 'type': '[str]'}, - 'archives': {'key': 'archives', 'type': '[str]'}, - 'configuration': {'key': 'conf', 'type': '{str}'}, - 'driver_memory': {'key': 'driverMemory', 'type': 'str'}, - 'driver_cores': {'key': 'driverCores', 'type': 'int'}, - 'executor_memory': {'key': 'executorMemory', 'type': 'str'}, - 'executor_cores': {'key': 'executorCores', 'type': 'int'}, - 'executor_count': {'key': 'numExecutors', 'type': 'int'}, - } - - def __init__( - self, - *, - name: Optional[str] = None, - file: Optional[str] = None, - class_name: Optional[str] = None, - arguments: Optional[List[str]] = None, - jars: Optional[List[str]] = None, - python_files: Optional[List[str]] = None, - files: Optional[List[str]] = None, - archives: Optional[List[str]] = None, - configuration: Optional[Dict[str, str]] = None, - driver_memory: Optional[str] = None, - driver_cores: Optional[int] = None, - executor_memory: Optional[str] = None, - executor_cores: Optional[int] = None, - executor_count: Optional[int] = None, - **kwargs - ): - super(SparkRequest, self).__init__(**kwargs) - self.name = name - self.file = file - self.class_name = class_name - self.arguments = arguments - self.jars = jars - self.python_files = python_files - self.files = files - self.archives = archives - self.configuration = configuration - self.driver_memory = driver_memory - self.driver_cores = driver_cores - self.executor_memory = executor_memory - self.executor_cores = executor_cores - self.executor_count = executor_count - - -class SparkScheduler(msrest.serialization.Model): - """SparkScheduler. - - :param submitted_at: - :type submitted_at: ~datetime.datetime - :param scheduled_at: - :type scheduled_at: ~datetime.datetime - :param ended_at: - :type ended_at: ~datetime.datetime - :param cancellation_requested_at: - :type cancellation_requested_at: ~datetime.datetime - :param current_state: Possible values include: "Queued", "Scheduled", "Ended". - :type current_state: str or ~azure.synapse.artifacts.v2020_12_01.models.SchedulerCurrentState - """ - - _attribute_map = { - 'submitted_at': {'key': 'submittedAt', 'type': 'iso-8601'}, - 'scheduled_at': {'key': 'scheduledAt', 'type': 'iso-8601'}, - 'ended_at': {'key': 'endedAt', 'type': 'iso-8601'}, - 'cancellation_requested_at': {'key': 'cancellationRequestedAt', 'type': 'iso-8601'}, - 'current_state': {'key': 'currentState', 'type': 'str'}, - } - - def __init__( - self, - *, - submitted_at: Optional[datetime.datetime] = None, - scheduled_at: Optional[datetime.datetime] = None, - ended_at: Optional[datetime.datetime] = None, - cancellation_requested_at: Optional[datetime.datetime] = None, - current_state: Optional[Union[str, "SchedulerCurrentState"]] = None, - **kwargs - ): - super(SparkScheduler, self).__init__(**kwargs) - self.submitted_at = submitted_at - self.scheduled_at = scheduled_at - self.ended_at = ended_at - self.cancellation_requested_at = cancellation_requested_at - self.current_state = current_state - - -class SparkServiceError(msrest.serialization.Model): - """SparkServiceError. - - :param message: - :type message: str - :param error_code: - :type error_code: str - :param source: Possible values include: "System", "User", "Unknown", "Dependency". - :type source: str or ~azure.synapse.artifacts.v2020_12_01.models.SparkErrorSource - """ - - _attribute_map = { - 'message': {'key': 'message', 'type': 'str'}, - 'error_code': {'key': 'errorCode', 'type': 'str'}, - 'source': {'key': 'source', 'type': 'str'}, - } - - def __init__( - self, - *, - message: Optional[str] = None, - error_code: Optional[str] = None, - source: Optional[Union[str, "SparkErrorSource"]] = None, - **kwargs - ): - super(SparkServiceError, self).__init__(**kwargs) - self.message = message - self.error_code = error_code - self.source = source - - -class SparkServicePlugin(msrest.serialization.Model): - """SparkServicePlugin. - - :param preparation_started_at: - :type preparation_started_at: ~datetime.datetime - :param resource_acquisition_started_at: - :type resource_acquisition_started_at: ~datetime.datetime - :param submission_started_at: - :type submission_started_at: ~datetime.datetime - :param monitoring_started_at: - :type monitoring_started_at: ~datetime.datetime - :param cleanup_started_at: - :type cleanup_started_at: ~datetime.datetime - :param current_state: Possible values include: "Preparation", "ResourceAcquisition", "Queued", - "Submission", "Monitoring", "Cleanup", "Ended". - :type current_state: str or ~azure.synapse.artifacts.v2020_12_01.models.PluginCurrentState - """ - - _attribute_map = { - 'preparation_started_at': {'key': 'preparationStartedAt', 'type': 'iso-8601'}, - 'resource_acquisition_started_at': {'key': 'resourceAcquisitionStartedAt', 'type': 'iso-8601'}, - 'submission_started_at': {'key': 'submissionStartedAt', 'type': 'iso-8601'}, - 'monitoring_started_at': {'key': 'monitoringStartedAt', 'type': 'iso-8601'}, - 'cleanup_started_at': {'key': 'cleanupStartedAt', 'type': 'iso-8601'}, - 'current_state': {'key': 'currentState', 'type': 'str'}, - } - - def __init__( - self, - *, - preparation_started_at: Optional[datetime.datetime] = None, - resource_acquisition_started_at: Optional[datetime.datetime] = None, - submission_started_at: Optional[datetime.datetime] = None, - monitoring_started_at: Optional[datetime.datetime] = None, - cleanup_started_at: Optional[datetime.datetime] = None, - current_state: Optional[Union[str, "PluginCurrentState"]] = None, - **kwargs - ): - super(SparkServicePlugin, self).__init__(**kwargs) - self.preparation_started_at = preparation_started_at - self.resource_acquisition_started_at = resource_acquisition_started_at - self.submission_started_at = submission_started_at - self.monitoring_started_at = monitoring_started_at - self.cleanup_started_at = cleanup_started_at - self.current_state = current_state - - -class SparkSource(TabularSource): - """A copy activity Spark Server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(SparkSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'SparkSource' # type: str - self.query = query - - -class SqlConnection(msrest.serialization.Model): - """The connection used to execute the SQL script. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The type of the connection. Possible values include: "SqlOnDemand", - "SqlPool". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.SqlConnectionType - :param name: Required. The identifier of the connection. - :type name: str - """ - - _validation = { - 'type': {'required': True}, - 'name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - } - - def __init__( - self, - *, - type: Union[str, "SqlConnectionType"], - name: str, - additional_properties: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(SqlConnection, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.type = type - self.name = name - - -class SqlDWSink(CopySink): - """A copy activity SQL Data Warehouse sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType - string). - :type pre_copy_script: any - :param allow_poly_base: Indicates to use PolyBase to copy data into SQL Data Warehouse when - applicable. Type: boolean (or Expression with resultType boolean). - :type allow_poly_base: any - :param poly_base_settings: Specifies PolyBase-related settings when allowPolyBase is true. - :type poly_base_settings: ~azure.synapse.artifacts.v2020_12_01.models.PolybaseSettings - :param allow_copy_command: Indicates to use Copy Command to copy data into SQL Data Warehouse. - Type: boolean (or Expression with resultType boolean). - :type allow_copy_command: any - :param copy_command_settings: Specifies Copy Command related settings when allowCopyCommand is - true. - :type copy_command_settings: ~azure.synapse.artifacts.v2020_12_01.models.DWCopyCommandSettings - :param table_option: The option to handle sink table, such as autoCreate. For now only - 'autoCreate' value is supported. Type: string (or Expression with resultType string). - :type table_option: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - 'allow_poly_base': {'key': 'allowPolyBase', 'type': 'object'}, - 'poly_base_settings': {'key': 'polyBaseSettings', 'type': 'PolybaseSettings'}, - 'allow_copy_command': {'key': 'allowCopyCommand', 'type': 'object'}, - 'copy_command_settings': {'key': 'copyCommandSettings', 'type': 'DWCopyCommandSettings'}, - 'table_option': {'key': 'tableOption', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - pre_copy_script: Optional[Any] = None, - allow_poly_base: Optional[Any] = None, - poly_base_settings: Optional["PolybaseSettings"] = None, - allow_copy_command: Optional[Any] = None, - copy_command_settings: Optional["DWCopyCommandSettings"] = None, - table_option: Optional[Any] = None, - **kwargs - ): - super(SqlDWSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'SqlDWSink' # type: str - self.pre_copy_script = pre_copy_script - self.allow_poly_base = allow_poly_base - self.poly_base_settings = poly_base_settings - self.allow_copy_command = allow_copy_command - self.copy_command_settings = copy_command_settings - self.table_option = table_option - - -class SqlDWSource(TabularSource): - """A copy activity SQL Data Warehouse source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param sql_reader_query: SQL Data Warehouse reader query. Type: string (or Expression with - resultType string). - :type sql_reader_query: any - :param sql_reader_stored_procedure_name: Name of the stored procedure for a SQL Data Warehouse - source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression - with resultType string). - :type sql_reader_stored_procedure_name: any - :param stored_procedure_parameters: Value and type setting for stored procedure parameters. - Example: "{Parameter1: {value: "1", type: "int"}}". Type: object (or Expression with resultType - object), itemType: StoredProcedureParameter. - :type stored_procedure_parameters: any - :param partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". - :type partition_option: any - :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.SqlPartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, - 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'object'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - sql_reader_query: Optional[Any] = None, - sql_reader_stored_procedure_name: Optional[Any] = None, - stored_procedure_parameters: Optional[Any] = None, - partition_option: Optional[Any] = None, - partition_settings: Optional["SqlPartitionSettings"] = None, - **kwargs - ): - super(SqlDWSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'SqlDWSource' # type: str - self.sql_reader_query = sql_reader_query - self.sql_reader_stored_procedure_name = sql_reader_stored_procedure_name - self.stored_procedure_parameters = stored_procedure_parameters - self.partition_option = partition_option - self.partition_settings = partition_settings - - -class SqlMISink(CopySink): - """A copy activity Azure SQL Managed Instance sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param sql_writer_stored_procedure_name: SQL writer stored procedure name. Type: string (or - Expression with resultType string). - :type sql_writer_stored_procedure_name: any - :param sql_writer_table_type: SQL writer table type. Type: string (or Expression with - resultType string). - :type sql_writer_table_type: any - :param pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType - string). - :type pre_copy_script: any - :param stored_procedure_parameters: SQL stored procedure parameters. - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - :param stored_procedure_table_type_parameter_name: The stored procedure parameter name of the - table type. Type: string (or Expression with resultType string). - :type stored_procedure_table_type_parameter_name: any - :param table_option: The option to handle sink table, such as autoCreate. For now only - 'autoCreate' value is supported. Type: string (or Expression with resultType string). - :type table_option: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'sql_writer_stored_procedure_name': {'key': 'sqlWriterStoredProcedureName', 'type': 'object'}, - 'sql_writer_table_type': {'key': 'sqlWriterTableType', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - 'stored_procedure_table_type_parameter_name': {'key': 'storedProcedureTableTypeParameterName', 'type': 'object'}, - 'table_option': {'key': 'tableOption', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - sql_writer_stored_procedure_name: Optional[Any] = None, - sql_writer_table_type: Optional[Any] = None, - pre_copy_script: Optional[Any] = None, - stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, - stored_procedure_table_type_parameter_name: Optional[Any] = None, - table_option: Optional[Any] = None, - **kwargs - ): - super(SqlMISink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'SqlMISink' # type: str - self.sql_writer_stored_procedure_name = sql_writer_stored_procedure_name - self.sql_writer_table_type = sql_writer_table_type - self.pre_copy_script = pre_copy_script - self.stored_procedure_parameters = stored_procedure_parameters - self.stored_procedure_table_type_parameter_name = stored_procedure_table_type_parameter_name - self.table_option = table_option - - -class SqlMISource(TabularSource): - """A copy activity Azure SQL Managed Instance source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). - :type sql_reader_query: any - :param sql_reader_stored_procedure_name: Name of the stored procedure for a Azure SQL Managed - Instance source. This cannot be used at the same time as SqlReaderQuery. Type: string (or - Expression with resultType string). - :type sql_reader_stored_procedure_name: any - :param stored_procedure_parameters: Value and type setting for stored procedure parameters. - Example: "{Parameter1: {value: "1", type: "int"}}". - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - :param produce_additional_types: Which additional types to produce. - :type produce_additional_types: any - :param partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". - :type partition_option: any - :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.SqlPartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, - 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - 'produce_additional_types': {'key': 'produceAdditionalTypes', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'object'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - sql_reader_query: Optional[Any] = None, - sql_reader_stored_procedure_name: Optional[Any] = None, - stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, - produce_additional_types: Optional[Any] = None, - partition_option: Optional[Any] = None, - partition_settings: Optional["SqlPartitionSettings"] = None, - **kwargs - ): - super(SqlMISource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'SqlMISource' # type: str - self.sql_reader_query = sql_reader_query - self.sql_reader_stored_procedure_name = sql_reader_stored_procedure_name - self.stored_procedure_parameters = stored_procedure_parameters - self.produce_additional_types = produce_additional_types - self.partition_option = partition_option - self.partition_settings = partition_settings - - -class SqlPartitionSettings(msrest.serialization.Model): - """The settings that will be leveraged for Sql source partitioning. - - :param partition_column_name: The name of the column in integer or datetime type that will be - used for proceeding partitioning. If not specified, the primary key of the table is - auto-detected and used as the partition column. Type: string (or Expression with resultType - string). - :type partition_column_name: any - :param partition_upper_bound: The maximum value of the partition column for partition range - splitting. This value is used to decide the partition stride, not for filtering the rows in - table. All rows in the table or query result will be partitioned and copied. Type: string (or - Expression with resultType string). - :type partition_upper_bound: any - :param partition_lower_bound: The minimum value of the partition column for partition range - splitting. This value is used to decide the partition stride, not for filtering the rows in - table. All rows in the table or query result will be partitioned and copied. Type: string (or - Expression with resultType string). - :type partition_lower_bound: any - """ - - _attribute_map = { - 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, - 'partition_upper_bound': {'key': 'partitionUpperBound', 'type': 'object'}, - 'partition_lower_bound': {'key': 'partitionLowerBound', 'type': 'object'}, - } - - def __init__( - self, - *, - partition_column_name: Optional[Any] = None, - partition_upper_bound: Optional[Any] = None, - partition_lower_bound: Optional[Any] = None, - **kwargs - ): - super(SqlPartitionSettings, self).__init__(**kwargs) - self.partition_column_name = partition_column_name - self.partition_upper_bound = partition_upper_bound - self.partition_lower_bound = partition_lower_bound - - -class SqlPool(TrackedResource): - """A SQL Analytics pool. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :param tags: A set of tags. Resource tags. - :type tags: dict[str, str] - :param location: Required. The geo-location where the resource lives. - :type location: str - :param sku: SQL pool SKU. - :type sku: ~azure.synapse.artifacts.v2020_12_01.models.Sku - :param max_size_bytes: Maximum size in bytes. - :type max_size_bytes: long - :param collation: Collation mode. - :type collation: str - :param source_database_id: Source database to create from. - :type source_database_id: str - :param recoverable_database_id: Backup database to restore from. - :type recoverable_database_id: str - :param provisioning_state: Resource state. - :type provisioning_state: str - :param status: Resource status. - :type status: str - :param restore_point_in_time: Snapshot time to restore. - :type restore_point_in_time: str - :param create_mode: What is this?. - :type create_mode: str - :param creation_date: Date the SQL pool was created. - :type creation_date: ~datetime.datetime - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'location': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'tags': {'key': 'tags', 'type': '{str}'}, - 'location': {'key': 'location', 'type': 'str'}, - 'sku': {'key': 'sku', 'type': 'Sku'}, - 'max_size_bytes': {'key': 'properties.maxSizeBytes', 'type': 'long'}, - 'collation': {'key': 'properties.collation', 'type': 'str'}, - 'source_database_id': {'key': 'properties.sourceDatabaseId', 'type': 'str'}, - 'recoverable_database_id': {'key': 'properties.recoverableDatabaseId', 'type': 'str'}, - 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, - 'status': {'key': 'properties.status', 'type': 'str'}, - 'restore_point_in_time': {'key': 'properties.restorePointInTime', 'type': 'str'}, - 'create_mode': {'key': 'properties.createMode', 'type': 'str'}, - 'creation_date': {'key': 'properties.creationDate', 'type': 'iso-8601'}, - } - - def __init__( - self, - *, - location: str, - tags: Optional[Dict[str, str]] = None, - sku: Optional["Sku"] = None, - max_size_bytes: Optional[int] = None, - collation: Optional[str] = None, - source_database_id: Optional[str] = None, - recoverable_database_id: Optional[str] = None, - provisioning_state: Optional[str] = None, - status: Optional[str] = None, - restore_point_in_time: Optional[str] = None, - create_mode: Optional[str] = None, - creation_date: Optional[datetime.datetime] = None, - **kwargs - ): - super(SqlPool, self).__init__(tags=tags, location=location, **kwargs) - self.sku = sku - self.max_size_bytes = max_size_bytes - self.collation = collation - self.source_database_id = source_database_id - self.recoverable_database_id = recoverable_database_id - self.provisioning_state = provisioning_state - self.status = status - self.restore_point_in_time = restore_point_in_time - self.create_mode = create_mode - self.creation_date = creation_date - - -class SqlPoolInfoListResult(msrest.serialization.Model): - """List of SQL pools. - - :param next_link: Link to the next page of results. - :type next_link: str - :param value: List of SQL pools. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.SqlPool] - """ - - _attribute_map = { - 'next_link': {'key': 'nextLink', 'type': 'str'}, - 'value': {'key': 'value', 'type': '[SqlPool]'}, - } - - def __init__( - self, - *, - next_link: Optional[str] = None, - value: Optional[List["SqlPool"]] = None, - **kwargs - ): - super(SqlPoolInfoListResult, self).__init__(**kwargs) - self.next_link = next_link - self.value = value - - -class SqlPoolReference(msrest.serialization.Model): - """SQL pool reference type. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. SQL pool reference type. Possible values include: "SqlPoolReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.SqlPoolReferenceType - :param reference_name: Required. Reference SQL pool name. - :type reference_name: str - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - } - - def __init__( - self, - *, - type: Union[str, "SqlPoolReferenceType"], - reference_name: str, - **kwargs - ): - super(SqlPoolReference, self).__init__(**kwargs) - self.type = type - self.reference_name = reference_name - - -class SqlPoolStoredProcedureActivity(Activity): - """Execute SQL pool stored procedure activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param sql_pool: Required. SQL pool stored procedure reference. - :type sql_pool: ~azure.synapse.artifacts.v2020_12_01.models.SqlPoolReference - :param stored_procedure_name: Required. Stored procedure name. Type: string (or Expression with - resultType string). - :type stored_procedure_name: any - :param stored_procedure_parameters: Value and type setting for stored procedure parameters. - Example: "{Parameter1: {value: "1", type: "int"}}". - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'sql_pool': {'required': True}, - 'stored_procedure_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'sql_pool': {'key': 'sqlPool', 'type': 'SqlPoolReference'}, - 'stored_procedure_name': {'key': 'typeProperties.storedProcedureName', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'typeProperties.storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - } - - def __init__( - self, - *, - name: str, - sql_pool: "SqlPoolReference", - stored_procedure_name: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, - **kwargs - ): - super(SqlPoolStoredProcedureActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) - self.type = 'SqlPoolStoredProcedure' # type: str - self.sql_pool = sql_pool - self.stored_procedure_name = stored_procedure_name - self.stored_procedure_parameters = stored_procedure_parameters - - -class SqlScript(msrest.serialization.Model): - """SQL script. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param description: The description of the SQL script. - :type description: str - :param type: The type of the SQL script. Possible values include: "SqlQuery". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.SqlScriptType - :param content: Required. The content of the SQL script. - :type content: ~azure.synapse.artifacts.v2020_12_01.models.SqlScriptContent - """ - - _validation = { - 'content': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'description', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'content': {'key': 'content', 'type': 'SqlScriptContent'}, - } - - def __init__( - self, - *, - content: "SqlScriptContent", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - type: Optional[Union[str, "SqlScriptType"]] = None, - **kwargs - ): - super(SqlScript, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.description = description - self.type = type - self.content = content - - -class SqlScriptContent(msrest.serialization.Model): - """The content of the SQL script. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param query: Required. SQL query to execute. - :type query: str - :param current_connection: Required. The connection used to execute the SQL script. - :type current_connection: ~azure.synapse.artifacts.v2020_12_01.models.SqlConnection - :param metadata: The metadata of the SQL script. - :type metadata: ~azure.synapse.artifacts.v2020_12_01.models.SqlScriptMetadata - """ - - _validation = { - 'query': {'required': True}, - 'current_connection': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'query': {'key': 'query', 'type': 'str'}, - 'current_connection': {'key': 'currentConnection', 'type': 'SqlConnection'}, - 'metadata': {'key': 'metadata', 'type': 'SqlScriptMetadata'}, - } - - def __init__( - self, - *, - query: str, - current_connection: "SqlConnection", - additional_properties: Optional[Dict[str, Any]] = None, - metadata: Optional["SqlScriptMetadata"] = None, - **kwargs - ): - super(SqlScriptContent, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.query = query - self.current_connection = current_connection - self.metadata = metadata - - -class SqlScriptMetadata(msrest.serialization.Model): - """The metadata of the SQL script. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param language: The language of the SQL script. - :type language: str - """ - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'language': {'key': 'language', 'type': 'str'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - language: Optional[str] = None, - **kwargs - ): - super(SqlScriptMetadata, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.language = language - - -class SqlScriptResource(msrest.serialization.Model): - """Sql Script resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource Id for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :param name: Required. The name of the resource. - :type name: str - :ivar type: The type of the resource. Ex- Microsoft.Compute/virtualMachines or - Microsoft.Storage/storageAccounts. - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Properties of sql script. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.SqlScript - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'required': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'SqlScript'}, - } - - def __init__( - self, - *, - name: str, - properties: "SqlScript", - **kwargs - ): - super(SqlScriptResource, self).__init__(**kwargs) - self.id = None - self.name = name - self.type = None - self.etag = None - self.properties = properties - - -class SqlScriptsListResponse(msrest.serialization.Model): - """A list of sql scripts resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of sql scripts. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.SqlScriptResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[SqlScriptResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - *, - value: List["SqlScriptResource"], - next_link: Optional[str] = None, - **kwargs - ): - super(SqlScriptsListResponse, self).__init__(**kwargs) - self.value = value - self.next_link = next_link - - -class SqlServerLinkedService(LinkedService): - """SQL Server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Required. The connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param user_name: The on-premises Windows authentication user name. Type: string (or Expression - with resultType string). - :type user_name: any - :param password: The on-premises Windows authentication password. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'connection_string': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'user_name': {'key': 'typeProperties.userName', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - connection_string: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - user_name: Optional[Any] = None, - password: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(SqlServerLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'SqlServer' # type: str - self.connection_string = connection_string - self.user_name = user_name - self.password = password - self.encrypted_credential = encrypted_credential - - -class SqlServerSink(CopySink): - """A copy activity SQL server sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param sql_writer_stored_procedure_name: SQL writer stored procedure name. Type: string (or - Expression with resultType string). - :type sql_writer_stored_procedure_name: any - :param sql_writer_table_type: SQL writer table type. Type: string (or Expression with - resultType string). - :type sql_writer_table_type: any - :param pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType - string). - :type pre_copy_script: any - :param stored_procedure_parameters: SQL stored procedure parameters. - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - :param stored_procedure_table_type_parameter_name: The stored procedure parameter name of the - table type. Type: string (or Expression with resultType string). - :type stored_procedure_table_type_parameter_name: any - :param table_option: The option to handle sink table, such as autoCreate. For now only - 'autoCreate' value is supported. Type: string (or Expression with resultType string). - :type table_option: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'sql_writer_stored_procedure_name': {'key': 'sqlWriterStoredProcedureName', 'type': 'object'}, - 'sql_writer_table_type': {'key': 'sqlWriterTableType', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - 'stored_procedure_table_type_parameter_name': {'key': 'storedProcedureTableTypeParameterName', 'type': 'object'}, - 'table_option': {'key': 'tableOption', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - sql_writer_stored_procedure_name: Optional[Any] = None, - sql_writer_table_type: Optional[Any] = None, - pre_copy_script: Optional[Any] = None, - stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, - stored_procedure_table_type_parameter_name: Optional[Any] = None, - table_option: Optional[Any] = None, - **kwargs - ): - super(SqlServerSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'SqlServerSink' # type: str - self.sql_writer_stored_procedure_name = sql_writer_stored_procedure_name - self.sql_writer_table_type = sql_writer_table_type - self.pre_copy_script = pre_copy_script - self.stored_procedure_parameters = stored_procedure_parameters - self.stored_procedure_table_type_parameter_name = stored_procedure_table_type_parameter_name - self.table_option = table_option - - -class SqlServerSource(TabularSource): - """A copy activity SQL server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). - :type sql_reader_query: any - :param sql_reader_stored_procedure_name: Name of the stored procedure for a SQL Database - source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression - with resultType string). - :type sql_reader_stored_procedure_name: any - :param stored_procedure_parameters: Value and type setting for stored procedure parameters. - Example: "{Parameter1: {value: "1", type: "int"}}". - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - :param produce_additional_types: Which additional types to produce. - :type produce_additional_types: any - :param partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". - :type partition_option: any - :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.SqlPartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, - 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - 'produce_additional_types': {'key': 'produceAdditionalTypes', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'object'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - sql_reader_query: Optional[Any] = None, - sql_reader_stored_procedure_name: Optional[Any] = None, - stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, - produce_additional_types: Optional[Any] = None, - partition_option: Optional[Any] = None, - partition_settings: Optional["SqlPartitionSettings"] = None, - **kwargs - ): - super(SqlServerSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'SqlServerSource' # type: str - self.sql_reader_query = sql_reader_query - self.sql_reader_stored_procedure_name = sql_reader_stored_procedure_name - self.stored_procedure_parameters = stored_procedure_parameters - self.produce_additional_types = produce_additional_types - self.partition_option = partition_option - self.partition_settings = partition_settings - - -class SqlServerStoredProcedureActivity(ExecutionActivity): - """SQL stored procedure activity type. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param stored_procedure_name: Required. Stored procedure name. Type: string (or Expression with - resultType string). - :type stored_procedure_name: any - :param stored_procedure_parameters: Value and type setting for stored procedure parameters. - Example: "{Parameter1: {value: "1", type: "int"}}". - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'stored_procedure_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'stored_procedure_name': {'key': 'typeProperties.storedProcedureName', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'typeProperties.storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - } - - def __init__( - self, - *, - name: str, - stored_procedure_name: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, - **kwargs - ): - super(SqlServerStoredProcedureActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'SqlServerStoredProcedure' # type: str - self.stored_procedure_name = stored_procedure_name - self.stored_procedure_parameters = stored_procedure_parameters - - -class SqlServerTableDataset(Dataset): - """The on-premises SQL Server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param schema_type_properties_schema: The schema name of the SQL Server dataset. Type: string - (or Expression with resultType string). - :type schema_type_properties_schema: any - :param table: The table name of the SQL Server dataset. Type: string (or Expression with - resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - schema_type_properties_schema: Optional[Any] = None, - table: Optional[Any] = None, - **kwargs - ): - super(SqlServerTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'SqlServerTable' # type: str - self.table_name = table_name - self.schema_type_properties_schema = schema_type_properties_schema - self.table = table - - -class SqlSink(CopySink): - """A copy activity SQL sink. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy sink type.Constant filled by server. - :type type: str - :param write_batch_size: Write batch size. Type: integer (or Expression with resultType - integer), minimum: 0. - :type write_batch_size: any - :param write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type write_batch_timeout: any - :param sink_retry_count: Sink retry count. Type: integer (or Expression with resultType - integer). - :type sink_retry_count: any - :param sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type sink_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the sink data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param sql_writer_stored_procedure_name: SQL writer stored procedure name. Type: string (or - Expression with resultType string). - :type sql_writer_stored_procedure_name: any - :param sql_writer_table_type: SQL writer table type. Type: string (or Expression with - resultType string). - :type sql_writer_table_type: any - :param pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType - string). - :type pre_copy_script: any - :param stored_procedure_parameters: SQL stored procedure parameters. - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - :param stored_procedure_table_type_parameter_name: The stored procedure parameter name of the - table type. Type: string (or Expression with resultType string). - :type stored_procedure_table_type_parameter_name: any - :param table_option: The option to handle sink table, such as autoCreate. For now only - 'autoCreate' value is supported. Type: string (or Expression with resultType string). - :type table_option: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, - 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, - 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, - 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'sql_writer_stored_procedure_name': {'key': 'sqlWriterStoredProcedureName', 'type': 'object'}, - 'sql_writer_table_type': {'key': 'sqlWriterTableType', 'type': 'object'}, - 'pre_copy_script': {'key': 'preCopyScript', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - 'stored_procedure_table_type_parameter_name': {'key': 'storedProcedureTableTypeParameterName', 'type': 'object'}, - 'table_option': {'key': 'tableOption', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - write_batch_size: Optional[Any] = None, - write_batch_timeout: Optional[Any] = None, - sink_retry_count: Optional[Any] = None, - sink_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - sql_writer_stored_procedure_name: Optional[Any] = None, - sql_writer_table_type: Optional[Any] = None, - pre_copy_script: Optional[Any] = None, - stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, - stored_procedure_table_type_parameter_name: Optional[Any] = None, - table_option: Optional[Any] = None, - **kwargs - ): - super(SqlSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'SqlSink' # type: str - self.sql_writer_stored_procedure_name = sql_writer_stored_procedure_name - self.sql_writer_table_type = sql_writer_table_type - self.pre_copy_script = pre_copy_script - self.stored_procedure_parameters = stored_procedure_parameters - self.stored_procedure_table_type_parameter_name = stored_procedure_table_type_parameter_name - self.table_option = table_option - - -class SqlSource(TabularSource): - """A copy activity SQL source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). - :type sql_reader_query: any - :param sql_reader_stored_procedure_name: Name of the stored procedure for a SQL Database - source. This cannot be used at the same time as SqlReaderQuery. Type: string (or Expression - with resultType string). - :type sql_reader_stored_procedure_name: any - :param stored_procedure_parameters: Value and type setting for stored procedure parameters. - Example: "{Parameter1: {value: "1", type: "int"}}". - :type stored_procedure_parameters: dict[str, - ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameter] - :param isolation_level: Specifies the transaction locking behavior for the SQL source. Allowed - values: ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value - is ReadCommitted. Type: string (or Expression with resultType string). - :type isolation_level: any - :param partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". - :type partition_option: any - :param partition_settings: The settings that will be leveraged for Sql source partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.SqlPartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, - 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, - 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, - 'isolation_level': {'key': 'isolationLevel', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'object'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - sql_reader_query: Optional[Any] = None, - sql_reader_stored_procedure_name: Optional[Any] = None, - stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, - isolation_level: Optional[Any] = None, - partition_option: Optional[Any] = None, - partition_settings: Optional["SqlPartitionSettings"] = None, - **kwargs - ): - super(SqlSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'SqlSource' # type: str - self.sql_reader_query = sql_reader_query - self.sql_reader_stored_procedure_name = sql_reader_stored_procedure_name - self.stored_procedure_parameters = stored_procedure_parameters - self.isolation_level = isolation_level - self.partition_option = partition_option - self.partition_settings = partition_settings - - -class SquareLinkedService(LinkedService): - """Square Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_properties: Properties used to connect to Square. It is mutually exclusive - with any other properties in the linked service. Type: object. - :type connection_properties: any - :param host: Required. The URL of the Square instance. (i.e. mystore.mysquare.com). - :type host: any - :param client_id: Required. The client ID associated with your Square application. - :type client_id: any - :param client_secret: The client secret associated with your Square application. - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param redirect_uri: Required. The redirect URL assigned in the Square application dashboard. - (i.e. http://localhost:2500). - :type redirect_uri: any - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - 'client_id': {'required': True}, - 'redirect_uri': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_properties': {'key': 'typeProperties.connectionProperties', 'type': 'object'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, - 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, - 'redirect_uri': {'key': 'typeProperties.redirectUri', 'type': 'object'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - host: Any, - client_id: Any, - redirect_uri: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_properties: Optional[Any] = None, - client_secret: Optional["SecretBase"] = None, - use_encrypted_endpoints: Optional[Any] = None, - use_host_verification: Optional[Any] = None, - use_peer_verification: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(SquareLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Square' # type: str - self.connection_properties = connection_properties - self.host = host - self.client_id = client_id - self.client_secret = client_secret - self.redirect_uri = redirect_uri - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential - - -class SquareObjectDataset(Dataset): - """Square Service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(SquareObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'SquareObject' # type: str - self.table_name = table_name - - -class SquareSource(TabularSource): - """A copy activity Square Service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(SquareSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'SquareSource' # type: str - self.query = query - - -class SSISAccessCredential(msrest.serialization.Model): - """SSIS access credential. - - All required parameters must be populated in order to send to Azure. - - :param domain: Required. Domain for windows authentication. - :type domain: any - :param user_name: Required. UseName for windows authentication. - :type user_name: any - :param password: Required. Password for windows authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - """ - - _validation = { - 'domain': {'required': True}, - 'user_name': {'required': True}, - 'password': {'required': True}, - } - - _attribute_map = { - 'domain': {'key': 'domain', 'type': 'object'}, - 'user_name': {'key': 'userName', 'type': 'object'}, - 'password': {'key': 'password', 'type': 'SecretBase'}, - } - - def __init__( - self, - *, - domain: Any, - user_name: Any, - password: "SecretBase", - **kwargs - ): - super(SSISAccessCredential, self).__init__(**kwargs) - self.domain = domain - self.user_name = user_name - self.password = password - - -class SSISChildPackage(msrest.serialization.Model): - """SSIS embedded child package. - - All required parameters must be populated in order to send to Azure. - - :param package_path: Required. Path for embedded child package. Type: string (or Expression - with resultType string). - :type package_path: any - :param package_name: Name for embedded child package. - :type package_name: str - :param package_content: Required. Content for embedded child package. Type: string (or - Expression with resultType string). - :type package_content: any - :param package_last_modified_date: Last modified date for embedded child package. - :type package_last_modified_date: str - """ - - _validation = { - 'package_path': {'required': True}, - 'package_content': {'required': True}, - } - - _attribute_map = { - 'package_path': {'key': 'packagePath', 'type': 'object'}, - 'package_name': {'key': 'packageName', 'type': 'str'}, - 'package_content': {'key': 'packageContent', 'type': 'object'}, - 'package_last_modified_date': {'key': 'packageLastModifiedDate', 'type': 'str'}, - } - - def __init__( - self, - *, - package_path: Any, - package_content: Any, - package_name: Optional[str] = None, - package_last_modified_date: Optional[str] = None, - **kwargs - ): - super(SSISChildPackage, self).__init__(**kwargs) - self.package_path = package_path - self.package_name = package_name - self.package_content = package_content - self.package_last_modified_date = package_last_modified_date - - -class SSISExecutionCredential(msrest.serialization.Model): - """SSIS package execution credential. - - All required parameters must be populated in order to send to Azure. - - :param domain: Required. Domain for windows authentication. - :type domain: any - :param user_name: Required. UseName for windows authentication. - :type user_name: any - :param password: Required. Password for windows authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecureString - """ - - _validation = { - 'domain': {'required': True}, - 'user_name': {'required': True}, - 'password': {'required': True}, - } - - _attribute_map = { - 'domain': {'key': 'domain', 'type': 'object'}, - 'user_name': {'key': 'userName', 'type': 'object'}, - 'password': {'key': 'password', 'type': 'SecureString'}, - } - - def __init__( - self, - *, - domain: Any, - user_name: Any, - password: "SecureString", - **kwargs - ): - super(SSISExecutionCredential, self).__init__(**kwargs) - self.domain = domain - self.user_name = user_name - self.password = password - - -class SSISExecutionParameter(msrest.serialization.Model): - """SSIS execution parameter. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. SSIS package execution parameter value. Type: string (or Expression - with resultType string). - :type value: any - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': 'object'}, - } - - def __init__( - self, - *, - value: Any, - **kwargs - ): - super(SSISExecutionParameter, self).__init__(**kwargs) - self.value = value - - -class SSISLogLocation(msrest.serialization.Model): - """SSIS package execution log location. - - All required parameters must be populated in order to send to Azure. - - :param log_path: Required. The SSIS package execution log path. Type: string (or Expression - with resultType string). - :type log_path: any - :param type: Required. The type of SSIS log location. Possible values include: "File". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.SsisLogLocationType - :param access_credential: The package execution log access credential. - :type access_credential: ~azure.synapse.artifacts.v2020_12_01.models.SSISAccessCredential - :param log_refresh_interval: Specifies the interval to refresh log. The default interval is 5 - minutes. Type: string (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type log_refresh_interval: any - """ - - _validation = { - 'log_path': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'log_path': {'key': 'logPath', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'str'}, - 'access_credential': {'key': 'typeProperties.accessCredential', 'type': 'SSISAccessCredential'}, - 'log_refresh_interval': {'key': 'typeProperties.logRefreshInterval', 'type': 'object'}, - } - - def __init__( - self, - *, - log_path: Any, - type: Union[str, "SsisLogLocationType"], - access_credential: Optional["SSISAccessCredential"] = None, - log_refresh_interval: Optional[Any] = None, - **kwargs - ): - super(SSISLogLocation, self).__init__(**kwargs) - self.log_path = log_path - self.type = type - self.access_credential = access_credential - self.log_refresh_interval = log_refresh_interval - - -class SsisObjectMetadataStatusResponse(msrest.serialization.Model): - """The status of the operation. - - :param status: The status of the operation. - :type status: str - :param name: The operation name. - :type name: str - :param properties: The operation properties. - :type properties: str - :param error: The operation error message. - :type error: str - """ - - _attribute_map = { - 'status': {'key': 'status', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'str'}, - 'error': {'key': 'error', 'type': 'str'}, - } - - def __init__( - self, - *, - status: Optional[str] = None, - name: Optional[str] = None, - properties: Optional[str] = None, - error: Optional[str] = None, - **kwargs - ): - super(SsisObjectMetadataStatusResponse, self).__init__(**kwargs) - self.status = status - self.name = name - self.properties = properties - self.error = error - - -class SSISPackageLocation(msrest.serialization.Model): - """SSIS package location. - - :param package_path: The SSIS package path. Type: string (or Expression with resultType - string). - :type package_path: any - :param type: The type of SSIS package location. Possible values include: "SSISDB", "File", - "InlinePackage", "PackageStore". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.SsisPackageLocationType - :param package_password: Password of the package. - :type package_password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param access_credential: The package access credential. - :type access_credential: ~azure.synapse.artifacts.v2020_12_01.models.SSISAccessCredential - :param configuration_path: The configuration file of the package execution. Type: string (or - Expression with resultType string). - :type configuration_path: any - :param configuration_access_credential: The configuration file access credential. - :type configuration_access_credential: - ~azure.synapse.artifacts.v2020_12_01.models.SSISAccessCredential - :param package_name: The package name. - :type package_name: str - :param package_content: The embedded package content. Type: string (or Expression with - resultType string). - :type package_content: any - :param package_last_modified_date: The embedded package last modified date. - :type package_last_modified_date: str - :param child_packages: The embedded child package list. - :type child_packages: list[~azure.synapse.artifacts.v2020_12_01.models.SSISChildPackage] - """ - - _attribute_map = { - 'package_path': {'key': 'packagePath', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'str'}, - 'package_password': {'key': 'typeProperties.packagePassword', 'type': 'SecretBase'}, - 'access_credential': {'key': 'typeProperties.accessCredential', 'type': 'SSISAccessCredential'}, - 'configuration_path': {'key': 'typeProperties.configurationPath', 'type': 'object'}, - 'configuration_access_credential': {'key': 'typeProperties.configurationAccessCredential', 'type': 'SSISAccessCredential'}, - 'package_name': {'key': 'typeProperties.packageName', 'type': 'str'}, - 'package_content': {'key': 'typeProperties.packageContent', 'type': 'object'}, - 'package_last_modified_date': {'key': 'typeProperties.packageLastModifiedDate', 'type': 'str'}, - 'child_packages': {'key': 'typeProperties.childPackages', 'type': '[SSISChildPackage]'}, - } - - def __init__( - self, - *, - package_path: Optional[Any] = None, - type: Optional[Union[str, "SsisPackageLocationType"]] = None, - package_password: Optional["SecretBase"] = None, - access_credential: Optional["SSISAccessCredential"] = None, - configuration_path: Optional[Any] = None, - configuration_access_credential: Optional["SSISAccessCredential"] = None, - package_name: Optional[str] = None, - package_content: Optional[Any] = None, - package_last_modified_date: Optional[str] = None, - child_packages: Optional[List["SSISChildPackage"]] = None, - **kwargs - ): - super(SSISPackageLocation, self).__init__(**kwargs) - self.package_path = package_path - self.type = type - self.package_password = package_password - self.access_credential = access_credential - self.configuration_path = configuration_path - self.configuration_access_credential = configuration_access_credential - self.package_name = package_name - self.package_content = package_content - self.package_last_modified_date = package_last_modified_date - self.child_packages = child_packages - - -class SSISPropertyOverride(msrest.serialization.Model): - """SSIS property override. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. SSIS package property override value. Type: string (or Expression with - resultType string). - :type value: any - :param is_sensitive: Whether SSIS package property override value is sensitive data. Value will - be encrypted in SSISDB if it is true. - :type is_sensitive: bool - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': 'object'}, - 'is_sensitive': {'key': 'isSensitive', 'type': 'bool'}, - } - - def __init__( - self, - *, - value: Any, - is_sensitive: Optional[bool] = None, - **kwargs - ): - super(SSISPropertyOverride, self).__init__(**kwargs) - self.value = value - self.is_sensitive = is_sensitive - - -class StagingSettings(msrest.serialization.Model): - """Staging settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param linked_service_name: Required. Staging linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param path: The path to storage for storing the interim data. Type: string (or Expression with - resultType string). - :type path: any - :param enable_compression: Specifies whether to use compression when copying data via an - interim staging. Default value is false. Type: boolean (or Expression with resultType boolean). - :type enable_compression: any - """ - - _validation = { - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'path': {'key': 'path', 'type': 'object'}, - 'enable_compression': {'key': 'enableCompression', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - path: Optional[Any] = None, - enable_compression: Optional[Any] = None, - **kwargs - ): - super(StagingSettings, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.linked_service_name = linked_service_name - self.path = path - self.enable_compression = enable_compression - - -class StartDataFlowDebugSessionRequest(msrest.serialization.Model): - """Request body structure for starting data flow debug session. - - :param session_id: The ID of data flow debug session. - :type session_id: str - :param data_flow: Data flow instance. - :type data_flow: ~azure.synapse.artifacts.v2020_12_01.models.DataFlowResource - :param datasets: List of datasets. - :type datasets: list[~azure.synapse.artifacts.v2020_12_01.models.DatasetResource] - :param linked_services: List of linked services. - :type linked_services: list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceResource] - :param staging: Staging info for debug session. - :type staging: any - :param debug_settings: Data flow debug settings. - :type debug_settings: any - :param incremental_debug: The type of new Databricks cluster. - :type incremental_debug: bool - """ - - _attribute_map = { - 'session_id': {'key': 'sessionId', 'type': 'str'}, - 'data_flow': {'key': 'dataFlow', 'type': 'DataFlowResource'}, - 'datasets': {'key': 'datasets', 'type': '[DatasetResource]'}, - 'linked_services': {'key': 'linkedServices', 'type': '[LinkedServiceResource]'}, - 'staging': {'key': 'staging', 'type': 'object'}, - 'debug_settings': {'key': 'debugSettings', 'type': 'object'}, - 'incremental_debug': {'key': 'incrementalDebug', 'type': 'bool'}, - } - - def __init__( - self, - *, - session_id: Optional[str] = None, - data_flow: Optional["DataFlowResource"] = None, - datasets: Optional[List["DatasetResource"]] = None, - linked_services: Optional[List["LinkedServiceResource"]] = None, - staging: Optional[Any] = None, - debug_settings: Optional[Any] = None, - incremental_debug: Optional[bool] = None, - **kwargs - ): - super(StartDataFlowDebugSessionRequest, self).__init__(**kwargs) - self.session_id = session_id - self.data_flow = data_flow - self.datasets = datasets - self.linked_services = linked_services - self.staging = staging - self.debug_settings = debug_settings - self.incremental_debug = incremental_debug - - -class StartDataFlowDebugSessionResponse(msrest.serialization.Model): - """Response body structure for starting data flow debug session. - - :param job_version: The ID of data flow debug job version. - :type job_version: str - """ - - _attribute_map = { - 'job_version': {'key': 'jobVersion', 'type': 'str'}, - } - - def __init__( - self, - *, - job_version: Optional[str] = None, - **kwargs - ): - super(StartDataFlowDebugSessionResponse, self).__init__(**kwargs) - self.job_version = job_version - - -class StoredProcedureParameter(msrest.serialization.Model): - """SQL stored procedure parameter. - - :param value: Stored procedure parameter value. Type: string (or Expression with resultType - string). - :type value: any - :param type: Stored procedure parameter type. Possible values include: "String", "Int", - "Int64", "Decimal", "Guid", "Boolean", "Date". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.StoredProcedureParameterType - """ - - _attribute_map = { - 'value': {'key': 'value', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'str'}, - } - - def __init__( - self, - *, - value: Optional[Any] = None, - type: Optional[Union[str, "StoredProcedureParameterType"]] = None, - **kwargs - ): - super(StoredProcedureParameter, self).__init__(**kwargs) - self.value = value - self.type = type - - -class SwitchActivity(ControlActivity): - """This activity evaluates an expression and executes activities under the cases property that correspond to the expression evaluation expected in the equals property. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param on: Required. An expression that would evaluate to a string or integer. This is used to - determine the block of activities in cases that will be executed. - :type on: ~azure.synapse.artifacts.v2020_12_01.models.Expression - :param cases: List of cases that correspond to expected values of the 'on' property. This is an - optional property and if not provided, the activity will execute activities provided in - defaultActivities. - :type cases: list[~azure.synapse.artifacts.v2020_12_01.models.SwitchCase] - :param default_activities: List of activities to execute if no case condition is satisfied. - This is an optional property and if not provided, the activity will exit without any action. - :type default_activities: list[~azure.synapse.artifacts.v2020_12_01.models.Activity] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'on': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'on': {'key': 'typeProperties.on', 'type': 'Expression'}, - 'cases': {'key': 'typeProperties.cases', 'type': '[SwitchCase]'}, - 'default_activities': {'key': 'typeProperties.defaultActivities', 'type': '[Activity]'}, - } - - def __init__( - self, - *, - name: str, - on: "Expression", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - cases: Optional[List["SwitchCase"]] = None, - default_activities: Optional[List["Activity"]] = None, - **kwargs - ): - super(SwitchActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) - self.type = 'Switch' # type: str - self.on = on - self.cases = cases - self.default_activities = default_activities - - -class SwitchCase(msrest.serialization.Model): - """Switch cases with have a value and corresponding activities. - - :param value: Expected value that satisfies the expression result of the 'on' property. - :type value: str - :param activities: List of activities to execute for satisfied case condition. - :type activities: list[~azure.synapse.artifacts.v2020_12_01.models.Activity] - """ - - _attribute_map = { - 'value': {'key': 'value', 'type': 'str'}, - 'activities': {'key': 'activities', 'type': '[Activity]'}, - } - - def __init__( - self, - *, - value: Optional[str] = None, - activities: Optional[List["Activity"]] = None, - **kwargs - ): - super(SwitchCase, self).__init__(**kwargs) - self.value = value - self.activities = activities - - -class SybaseLinkedService(LinkedService): - """Linked service for Sybase data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param server: Required. Server name for connection. Type: string (or Expression with - resultType string). - :type server: any - :param database: Required. Database name for connection. Type: string (or Expression with - resultType string). - :type database: any - :param schema: Schema name for connection. Type: string (or Expression with resultType string). - :type schema: any - :param authentication_type: AuthenticationType to be used for connection. Possible values - include: "Basic", "Windows". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.SybaseAuthenticationType - :param username: Username for authentication. Type: string (or Expression with resultType - string). - :type username: any - :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'server': {'required': True}, - 'database': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'schema': {'key': 'typeProperties.schema', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - server: Any, - database: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - schema: Optional[Any] = None, - authentication_type: Optional[Union[str, "SybaseAuthenticationType"]] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(SybaseLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Sybase' # type: str - self.server = server - self.database = database - self.schema = schema - self.authentication_type = authentication_type - self.username = username - self.password = password - self.encrypted_credential = encrypted_credential - - -class SybaseSource(TabularSource): - """A copy activity source for Sybase databases. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Database query. Type: string (or Expression with resultType string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(SybaseSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'SybaseSource' # type: str - self.query = query - - -class SybaseTableDataset(Dataset): - """The Sybase table dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The Sybase table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(SybaseTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'SybaseTable' # type: str - self.table_name = table_name - - -class SynapseNotebookActivity(ExecutionActivity): - """Execute Synapse notebook activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param notebook: Required. Synapse notebook reference. - :type notebook: ~azure.synapse.artifacts.v2020_12_01.models.SynapseNotebookReference - :param parameters: Notebook parameters. - :type parameters: dict[str, any] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'notebook': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'notebook': {'key': 'typeProperties.notebook', 'type': 'SynapseNotebookReference'}, - 'parameters': {'key': 'typeProperties.parameters', 'type': '{object}'}, - } - - def __init__( - self, - *, - name: str, - notebook: "SynapseNotebookReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - parameters: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(SynapseNotebookActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'SynapseNotebook' # type: str - self.notebook = notebook - self.parameters = parameters - - -class SynapseNotebookReference(msrest.serialization.Model): - """Synapse notebook reference type. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Synapse notebook reference type. Possible values include: - "NotebookReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.NotebookReferenceType - :param reference_name: Required. Reference notebook name. - :type reference_name: str - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - } - - def __init__( - self, - *, - type: Union[str, "NotebookReferenceType"], - reference_name: str, - **kwargs - ): - super(SynapseNotebookReference, self).__init__(**kwargs) - self.type = type - self.reference_name = reference_name - - -class SynapseSparkJobDefinitionActivity(ExecutionActivity): - """Execute spark job activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param spark_job: Required. Synapse spark job reference. - :type spark_job: ~azure.synapse.artifacts.v2020_12_01.models.SynapseSparkJobReference - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'spark_job': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'spark_job': {'key': 'typeProperties.sparkJob', 'type': 'SynapseSparkJobReference'}, - } - - def __init__( - self, - *, - name: str, - spark_job: "SynapseSparkJobReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - **kwargs - ): - super(SynapseSparkJobDefinitionActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'SparkJob' # type: str - self.spark_job = spark_job - - -class SynapseSparkJobReference(msrest.serialization.Model): - """Synapse spark job reference type. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Synapse spark job reference type. Possible values include: - "SparkJobDefinitionReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.SparkJobReferenceType - :param reference_name: Required. Reference spark job name. - :type reference_name: str - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - } - - def __init__( - self, - *, - type: Union[str, "SparkJobReferenceType"], - reference_name: str, - **kwargs - ): - super(SynapseSparkJobReference, self).__init__(**kwargs) - self.type = type - self.reference_name = reference_name - - -class TabularTranslator(CopyTranslator): - """A copy activity tabular translator. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy translator type.Constant filled by server. - :type type: str - :param column_mappings: Column mappings. Example: "UserId: MyUserId, Group: MyGroup, Name: - MyName" Type: string (or Expression with resultType string). This property will be retired. - Please use mappings property. - :type column_mappings: any - :param schema_mapping: The schema mapping to map between tabular data and hierarchical data. - Example: {"Column1": "$.Column1", "Column2": "$.Column2.Property1", "Column3": - "$.Column2.Property2"}. Type: object (or Expression with resultType object). This property will - be retired. Please use mappings property. - :type schema_mapping: any - :param collection_reference: The JSON Path of the Nested Array that is going to do cross-apply. - Type: object (or Expression with resultType object). - :type collection_reference: any - :param map_complex_values_to_string: Whether to map complex (array and object) values to simple - strings in json format. Type: boolean (or Expression with resultType boolean). - :type map_complex_values_to_string: any - :param mappings: Column mappings with logical types. Tabular->tabular example: - [{"source":{"name":"CustomerName","type":"String"},"sink":{"name":"ClientName","type":"String"}},{"source":{"name":"CustomerAddress","type":"String"},"sink":{"name":"ClientAddress","type":"String"}}]. - Hierarchical->tabular example: - [{"source":{"path":"$.CustomerName","type":"String"},"sink":{"name":"ClientName","type":"String"}},{"source":{"path":"$.CustomerAddress","type":"String"},"sink":{"name":"ClientAddress","type":"String"}}]. - Type: object (or Expression with resultType object). - :type mappings: any - :param type_conversion: Whether to enable the advanced type conversion feature in the Copy - activity. Type: boolean (or Expression with resultType boolean). - :type type_conversion: any - :param type_conversion_settings: Type conversion settings. - :type type_conversion_settings: - ~azure.synapse.artifacts.v2020_12_01.models.TypeConversionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'column_mappings': {'key': 'columnMappings', 'type': 'object'}, - 'schema_mapping': {'key': 'schemaMapping', 'type': 'object'}, - 'collection_reference': {'key': 'collectionReference', 'type': 'object'}, - 'map_complex_values_to_string': {'key': 'mapComplexValuesToString', 'type': 'object'}, - 'mappings': {'key': 'mappings', 'type': 'object'}, - 'type_conversion': {'key': 'typeConversion', 'type': 'object'}, - 'type_conversion_settings': {'key': 'typeConversionSettings', 'type': 'TypeConversionSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - column_mappings: Optional[Any] = None, - schema_mapping: Optional[Any] = None, - collection_reference: Optional[Any] = None, - map_complex_values_to_string: Optional[Any] = None, - mappings: Optional[Any] = None, - type_conversion: Optional[Any] = None, - type_conversion_settings: Optional["TypeConversionSettings"] = None, - **kwargs - ): - super(TabularTranslator, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'TabularTranslator' # type: str - self.column_mappings = column_mappings - self.schema_mapping = schema_mapping - self.collection_reference = collection_reference - self.map_complex_values_to_string = map_complex_values_to_string - self.mappings = mappings - self.type_conversion = type_conversion - self.type_conversion_settings = type_conversion_settings - - -class TarGZipReadSettings(CompressionReadSettings): - """The TarGZip compression read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The Compression setting type.Constant filled by server. - :type type: str - :param preserve_compression_file_name_as_folder: Preserve the compression file name as folder - path. Type: boolean (or Expression with resultType boolean). - :type preserve_compression_file_name_as_folder: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'preserve_compression_file_name_as_folder': {'key': 'preserveCompressionFileNameAsFolder', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - preserve_compression_file_name_as_folder: Optional[Any] = None, - **kwargs - ): - super(TarGZipReadSettings, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'TarGZipReadSettings' # type: str - self.preserve_compression_file_name_as_folder = preserve_compression_file_name_as_folder - - -class TarReadSettings(CompressionReadSettings): - """The Tar compression read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The Compression setting type.Constant filled by server. - :type type: str - :param preserve_compression_file_name_as_folder: Preserve the compression file name as folder - path. Type: boolean (or Expression with resultType boolean). - :type preserve_compression_file_name_as_folder: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'preserve_compression_file_name_as_folder': {'key': 'preserveCompressionFileNameAsFolder', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - preserve_compression_file_name_as_folder: Optional[Any] = None, - **kwargs - ): - super(TarReadSettings, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'TarReadSettings' # type: str - self.preserve_compression_file_name_as_folder = preserve_compression_file_name_as_folder - - -class TeradataLinkedService(LinkedService): - """Linked service for Teradata data source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: Teradata ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param server: Server name for connection. Type: string (or Expression with resultType string). - :type server: any - :param authentication_type: AuthenticationType to be used for connection. Possible values - include: "Basic", "Windows". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.TeradataAuthenticationType - :param username: Username for authentication. Type: string (or Expression with resultType - string). - :type username: any - :param password: Password for authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'server': {'key': 'typeProperties.server', 'type': 'object'}, - 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, - 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_string: Optional[Any] = None, - server: Optional[Any] = None, - authentication_type: Optional[Union[str, "TeradataAuthenticationType"]] = None, - username: Optional[Any] = None, - password: Optional["SecretBase"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(TeradataLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Teradata' # type: str - self.connection_string = connection_string - self.server = server - self.authentication_type = authentication_type - self.username = username - self.password = password - self.encrypted_credential = encrypted_credential - - -class TeradataPartitionSettings(msrest.serialization.Model): - """The settings that will be leveraged for teradata source partitioning. - - :param partition_column_name: The name of the column that will be used for proceeding range or - hash partitioning. Type: string (or Expression with resultType string). - :type partition_column_name: any - :param partition_upper_bound: The maximum value of column specified in partitionColumnName that - will be used for proceeding range partitioning. Type: string (or Expression with resultType - string). - :type partition_upper_bound: any - :param partition_lower_bound: The minimum value of column specified in partitionColumnName that - will be used for proceeding range partitioning. Type: string (or Expression with resultType - string). - :type partition_lower_bound: any - """ - - _attribute_map = { - 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, - 'partition_upper_bound': {'key': 'partitionUpperBound', 'type': 'object'}, - 'partition_lower_bound': {'key': 'partitionLowerBound', 'type': 'object'}, - } - - def __init__( - self, - *, - partition_column_name: Optional[Any] = None, - partition_upper_bound: Optional[Any] = None, - partition_lower_bound: Optional[Any] = None, - **kwargs - ): - super(TeradataPartitionSettings, self).__init__(**kwargs) - self.partition_column_name = partition_column_name - self.partition_upper_bound = partition_upper_bound - self.partition_lower_bound = partition_lower_bound - - -class TeradataSource(TabularSource): - """A copy activity Teradata source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: Teradata query. Type: string (or Expression with resultType string). - :type query: any - :param partition_option: The partition mechanism that will be used for teradata read in - parallel. Possible values include: "None", "Hash", "DynamicRange". - :type partition_option: str or - ~azure.synapse.artifacts.v2020_12_01.models.TeradataPartitionOption - :param partition_settings: The settings that will be leveraged for teradata source - partitioning. - :type partition_settings: ~azure.synapse.artifacts.v2020_12_01.models.TeradataPartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'TeradataPartitionSettings'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - partition_option: Optional[Union[str, "TeradataPartitionOption"]] = None, - partition_settings: Optional["TeradataPartitionSettings"] = None, - **kwargs - ): - super(TeradataSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'TeradataSource' # type: str - self.query = query - self.partition_option = partition_option - self.partition_settings = partition_settings - - -class TeradataTableDataset(Dataset): - """The Teradata database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param database: The database name of Teradata. Type: string (or Expression with resultType - string). - :type database: any - :param table: The table name of Teradata. Type: string (or Expression with resultType string). - :type table: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - database: Optional[Any] = None, - table: Optional[Any] = None, - **kwargs - ): - super(TeradataTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'TeradataTable' # type: str - self.database = database - self.table = table - - -class TextFormat(DatasetStorageFormat): - """The data stored in text format. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset storage format.Constant filled by server. - :type type: str - :param serializer: Serializer. Type: string (or Expression with resultType string). - :type serializer: any - :param deserializer: Deserializer. Type: string (or Expression with resultType string). - :type deserializer: any - :param column_delimiter: The column delimiter. Type: string (or Expression with resultType - string). - :type column_delimiter: any - :param row_delimiter: The row delimiter. Type: string (or Expression with resultType string). - :type row_delimiter: any - :param escape_char: The escape character. Type: string (or Expression with resultType string). - :type escape_char: any - :param quote_char: The quote character. Type: string (or Expression with resultType string). - :type quote_char: any - :param null_value: The null value string. Type: string (or Expression with resultType string). - :type null_value: any - :param encoding_name: The code page name of the preferred encoding. If miss, the default value - is ΓÇ£utf-8ΓÇ¥, unless BOM denotes another Unicode encoding. Refer to the ΓÇ£NameΓÇ¥ column of - the table in the following link to set supported values: - https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string (or Expression with - resultType string). - :type encoding_name: any - :param treat_empty_as_null: Treat empty column values in the text file as null. The default - value is true. Type: boolean (or Expression with resultType boolean). - :type treat_empty_as_null: any - :param skip_line_count: The number of lines/rows to be skipped when parsing text files. The - default value is 0. Type: integer (or Expression with resultType integer). - :type skip_line_count: any - :param first_row_as_header: When used as input, treat the first row of data as headers. When - used as output,write the headers into the output as the first row of data. The default value is - false. Type: boolean (or Expression with resultType boolean). - :type first_row_as_header: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'serializer': {'key': 'serializer', 'type': 'object'}, - 'deserializer': {'key': 'deserializer', 'type': 'object'}, - 'column_delimiter': {'key': 'columnDelimiter', 'type': 'object'}, - 'row_delimiter': {'key': 'rowDelimiter', 'type': 'object'}, - 'escape_char': {'key': 'escapeChar', 'type': 'object'}, - 'quote_char': {'key': 'quoteChar', 'type': 'object'}, - 'null_value': {'key': 'nullValue', 'type': 'object'}, - 'encoding_name': {'key': 'encodingName', 'type': 'object'}, - 'treat_empty_as_null': {'key': 'treatEmptyAsNull', 'type': 'object'}, - 'skip_line_count': {'key': 'skipLineCount', 'type': 'object'}, - 'first_row_as_header': {'key': 'firstRowAsHeader', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - serializer: Optional[Any] = None, - deserializer: Optional[Any] = None, - column_delimiter: Optional[Any] = None, - row_delimiter: Optional[Any] = None, - escape_char: Optional[Any] = None, - quote_char: Optional[Any] = None, - null_value: Optional[Any] = None, - encoding_name: Optional[Any] = None, - treat_empty_as_null: Optional[Any] = None, - skip_line_count: Optional[Any] = None, - first_row_as_header: Optional[Any] = None, - **kwargs - ): - super(TextFormat, self).__init__(additional_properties=additional_properties, serializer=serializer, deserializer=deserializer, **kwargs) - self.type = 'TextFormat' # type: str - self.column_delimiter = column_delimiter - self.row_delimiter = row_delimiter - self.escape_char = escape_char - self.quote_char = quote_char - self.null_value = null_value - self.encoding_name = encoding_name - self.treat_empty_as_null = treat_empty_as_null - self.skip_line_count = skip_line_count - self.first_row_as_header = first_row_as_header - - -class TriggerDependencyProvisioningStatus(msrest.serialization.Model): - """Defines the response of a provision trigger dependency operation. - - All required parameters must be populated in order to send to Azure. - - :param trigger_name: Required. Trigger name. - :type trigger_name: str - :param provisioning_status: Required. Provisioning status. - :type provisioning_status: str - """ - - _validation = { - 'trigger_name': {'required': True}, - 'provisioning_status': {'required': True}, - } - - _attribute_map = { - 'trigger_name': {'key': 'triggerName', 'type': 'str'}, - 'provisioning_status': {'key': 'provisioningStatus', 'type': 'str'}, - } - - def __init__( - self, - *, - trigger_name: str, - provisioning_status: str, - **kwargs - ): - super(TriggerDependencyProvisioningStatus, self).__init__(**kwargs) - self.trigger_name = trigger_name - self.provisioning_status = provisioning_status - - -class TriggerDependencyReference(DependencyReference): - """Trigger referenced dependency. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: TumblingWindowTriggerDependencyReference. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. The type of dependency reference.Constant filled by server. - :type type: str - :param reference_trigger: Required. Referenced trigger. - :type reference_trigger: ~azure.synapse.artifacts.v2020_12_01.models.TriggerReference - """ - - _validation = { - 'type': {'required': True}, - 'reference_trigger': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_trigger': {'key': 'referenceTrigger', 'type': 'TriggerReference'}, - } - - _subtype_map = { - 'type': {'TumblingWindowTriggerDependencyReference': 'TumblingWindowTriggerDependencyReference'} - } - - def __init__( - self, - *, - reference_trigger: "TriggerReference", - **kwargs - ): - super(TriggerDependencyReference, self).__init__(**kwargs) - self.type = 'TriggerDependencyReference' # type: str - self.reference_trigger = reference_trigger - - -class TriggerListResponse(msrest.serialization.Model): - """A list of trigger resources. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of triggers. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.TriggerResource] - :param next_link: The link to the next page of results, if any remaining results exist. - :type next_link: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[TriggerResource]'}, - 'next_link': {'key': 'nextLink', 'type': 'str'}, - } - - def __init__( - self, - *, - value: List["TriggerResource"], - next_link: Optional[str] = None, - **kwargs - ): - super(TriggerListResponse, self).__init__(**kwargs) - self.value = value - self.next_link = next_link - - -class TriggerPipelineReference(msrest.serialization.Model): - """Pipeline that needs to be triggered with the given parameters. - - :param pipeline_reference: Pipeline reference. - :type pipeline_reference: ~azure.synapse.artifacts.v2020_12_01.models.PipelineReference - :param parameters: Pipeline parameters. - :type parameters: dict[str, any] - """ - - _attribute_map = { - 'pipeline_reference': {'key': 'pipelineReference', 'type': 'PipelineReference'}, - 'parameters': {'key': 'parameters', 'type': '{object}'}, - } - - def __init__( - self, - *, - pipeline_reference: Optional["PipelineReference"] = None, - parameters: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(TriggerPipelineReference, self).__init__(**kwargs) - self.pipeline_reference = pipeline_reference - self.parameters = parameters - - -class TriggerReference(msrest.serialization.Model): - """Trigger reference type. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Trigger reference type. Possible values include: "TriggerReference". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerReferenceType - :param reference_name: Required. Reference trigger name. - :type reference_name: str - """ - - _validation = { - 'type': {'required': True}, - 'reference_name': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_name': {'key': 'referenceName', 'type': 'str'}, - } - - def __init__( - self, - *, - type: Union[str, "TriggerReferenceType"], - reference_name: str, - **kwargs - ): - super(TriggerReference, self).__init__(**kwargs) - self.type = type - self.reference_name = reference_name - - -class TriggerResource(SubResource): - """Trigger resource type. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :ivar etag: Resource Etag. - :vartype etag: str - :param properties: Required. Properties of the trigger. - :type properties: ~azure.synapse.artifacts.v2020_12_01.models.Trigger - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'etag': {'readonly': True}, - 'properties': {'required': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'etag': {'key': 'etag', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': 'Trigger'}, - } - - def __init__( - self, - *, - properties: "Trigger", - **kwargs - ): - super(TriggerResource, self).__init__(**kwargs) - self.properties = properties - - -class TriggerRun(msrest.serialization.Model): - """Trigger runs. - - Variables are only populated by the server, and will be ignored when sending a request. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :ivar trigger_run_id: Trigger run id. - :vartype trigger_run_id: str - :ivar trigger_name: Trigger name. - :vartype trigger_name: str - :ivar trigger_type: Trigger type. - :vartype trigger_type: str - :ivar trigger_run_timestamp: Trigger run start time. - :vartype trigger_run_timestamp: ~datetime.datetime - :ivar status: Trigger run status. Possible values include: "Succeeded", "Failed", "Inprogress". - :vartype status: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRunStatus - :ivar message: Trigger error message. - :vartype message: str - :ivar properties: List of property name and value related to trigger run. Name, value pair - depends on type of trigger. - :vartype properties: dict[str, str] - :ivar triggered_pipelines: List of pipeline name and run Id triggered by the trigger run. - :vartype triggered_pipelines: dict[str, str] - """ - - _validation = { - 'trigger_run_id': {'readonly': True}, - 'trigger_name': {'readonly': True}, - 'trigger_type': {'readonly': True}, - 'trigger_run_timestamp': {'readonly': True}, - 'status': {'readonly': True}, - 'message': {'readonly': True}, - 'properties': {'readonly': True}, - 'triggered_pipelines': {'readonly': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'trigger_run_id': {'key': 'triggerRunId', 'type': 'str'}, - 'trigger_name': {'key': 'triggerName', 'type': 'str'}, - 'trigger_type': {'key': 'triggerType', 'type': 'str'}, - 'trigger_run_timestamp': {'key': 'triggerRunTimestamp', 'type': 'iso-8601'}, - 'status': {'key': 'status', 'type': 'str'}, - 'message': {'key': 'message', 'type': 'str'}, - 'properties': {'key': 'properties', 'type': '{str}'}, - 'triggered_pipelines': {'key': 'triggeredPipelines', 'type': '{str}'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - **kwargs - ): - super(TriggerRun, self).__init__(**kwargs) - self.additional_properties = additional_properties - self.trigger_run_id = None - self.trigger_name = None - self.trigger_type = None - self.trigger_run_timestamp = None - self.status = None - self.message = None - self.properties = None - self.triggered_pipelines = None - - -class TriggerRunsQueryResponse(msrest.serialization.Model): - """A list of trigger runs. - - All required parameters must be populated in order to send to Azure. - - :param value: Required. List of trigger runs. - :type value: list[~azure.synapse.artifacts.v2020_12_01.models.TriggerRun] - :param continuation_token: The continuation token for getting the next page of results, if any - remaining results exist, null otherwise. - :type continuation_token: str - """ - - _validation = { - 'value': {'required': True}, - } - - _attribute_map = { - 'value': {'key': 'value', 'type': '[TriggerRun]'}, - 'continuation_token': {'key': 'continuationToken', 'type': 'str'}, - } - - def __init__( - self, - *, - value: List["TriggerRun"], - continuation_token: Optional[str] = None, - **kwargs - ): - super(TriggerRunsQueryResponse, self).__init__(**kwargs) - self.value = value - self.continuation_token = continuation_token - - -class TriggerSubscriptionOperationStatus(msrest.serialization.Model): - """Defines the response of a trigger subscription operation. - - Variables are only populated by the server, and will be ignored when sending a request. - - :ivar trigger_name: Trigger name. - :vartype trigger_name: str - :ivar status: Event Subscription Status. Possible values include: "Enabled", "Provisioning", - "Deprovisioning", "Disabled", "Unknown". - :vartype status: str or ~azure.synapse.artifacts.v2020_12_01.models.EventSubscriptionStatus - """ - - _validation = { - 'trigger_name': {'readonly': True}, - 'status': {'readonly': True}, - } - - _attribute_map = { - 'trigger_name': {'key': 'triggerName', 'type': 'str'}, - 'status': {'key': 'status', 'type': 'str'}, - } - - def __init__( - self, - **kwargs - ): - super(TriggerSubscriptionOperationStatus, self).__init__(**kwargs) - self.trigger_name = None - self.status = None - - -class TumblingWindowTrigger(Trigger): - """Trigger that schedules pipeline runs for all fixed time interval windows from a start time without gaps and also supports backfill scenarios (when start time is in the past). - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Trigger type.Constant filled by server. - :type type: str - :param description: Trigger description. - :type description: str - :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are - called on the Trigger. Possible values include: "Started", "Stopped", "Disabled". - :vartype runtime_state: str or ~azure.synapse.artifacts.v2020_12_01.models.TriggerRuntimeState - :param annotations: List of tags that can be used for describing the trigger. - :type annotations: list[any] - :param pipeline: Required. Pipeline for which runs are created when an event is fired for - trigger window that is ready. - :type pipeline: ~azure.synapse.artifacts.v2020_12_01.models.TriggerPipelineReference - :param frequency: Required. The frequency of the time windows. Possible values include: - "Minute", "Hour", "Month". - :type frequency: str or ~azure.synapse.artifacts.v2020_12_01.models.TumblingWindowFrequency - :param interval: Required. The interval of the time windows. The minimum interval allowed is 15 - Minutes. - :type interval: int - :param start_time: Required. The start time for the time period for the trigger during which - events are fired for windows that are ready. Only UTC time is currently supported. - :type start_time: ~datetime.datetime - :param end_time: The end time for the time period for the trigger during which events are fired - for windows that are ready. Only UTC time is currently supported. - :type end_time: ~datetime.datetime - :param delay: Specifies how long the trigger waits past due time before triggering new run. It - doesn't alter window start and end time. The default is 0. Type: string (or Expression with - resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type delay: any - :param max_concurrency: Required. The max number of parallel time windows (ready for execution) - for which a new run is triggered. - :type max_concurrency: int - :param retry_policy: Retry policy that will be applied for failed pipeline runs. - :type retry_policy: ~azure.synapse.artifacts.v2020_12_01.models.RetryPolicy - :param depends_on: Triggers that this trigger depends on. Only tumbling window triggers are - supported. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.DependencyReference] - """ - - _validation = { - 'type': {'required': True}, - 'runtime_state': {'readonly': True}, - 'pipeline': {'required': True}, - 'frequency': {'required': True}, - 'interval': {'required': True}, - 'start_time': {'required': True}, - 'max_concurrency': {'required': True, 'maximum': 50, 'minimum': 1}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'pipeline': {'key': 'pipeline', 'type': 'TriggerPipelineReference'}, - 'frequency': {'key': 'typeProperties.frequency', 'type': 'str'}, - 'interval': {'key': 'typeProperties.interval', 'type': 'int'}, - 'start_time': {'key': 'typeProperties.startTime', 'type': 'iso-8601'}, - 'end_time': {'key': 'typeProperties.endTime', 'type': 'iso-8601'}, - 'delay': {'key': 'typeProperties.delay', 'type': 'object'}, - 'max_concurrency': {'key': 'typeProperties.maxConcurrency', 'type': 'int'}, - 'retry_policy': {'key': 'typeProperties.retryPolicy', 'type': 'RetryPolicy'}, - 'depends_on': {'key': 'typeProperties.dependsOn', 'type': '[DependencyReference]'}, - } - - def __init__( - self, - *, - pipeline: "TriggerPipelineReference", - frequency: Union[str, "TumblingWindowFrequency"], - interval: int, - start_time: datetime.datetime, - max_concurrency: int, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - annotations: Optional[List[Any]] = None, - end_time: Optional[datetime.datetime] = None, - delay: Optional[Any] = None, - retry_policy: Optional["RetryPolicy"] = None, - depends_on: Optional[List["DependencyReference"]] = None, - **kwargs - ): - super(TumblingWindowTrigger, self).__init__(additional_properties=additional_properties, description=description, annotations=annotations, **kwargs) - self.type = 'TumblingWindowTrigger' # type: str - self.pipeline = pipeline - self.frequency = frequency - self.interval = interval - self.start_time = start_time - self.end_time = end_time - self.delay = delay - self.max_concurrency = max_concurrency - self.retry_policy = retry_policy - self.depends_on = depends_on - - -class TumblingWindowTriggerDependencyReference(TriggerDependencyReference): - """Referenced tumbling window trigger dependency. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. The type of dependency reference.Constant filled by server. - :type type: str - :param reference_trigger: Required. Referenced trigger. - :type reference_trigger: ~azure.synapse.artifacts.v2020_12_01.models.TriggerReference - :param offset: Timespan applied to the start time of a tumbling window when evaluating - dependency. - :type offset: str - :param size: The size of the window when evaluating the dependency. If undefined the frequency - of the tumbling window will be used. - :type size: str - """ - - _validation = { - 'type': {'required': True}, - 'reference_trigger': {'required': True}, - 'offset': {'max_length': 15, 'min_length': 8, 'pattern': r'-?((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))'}, - 'size': {'max_length': 15, 'min_length': 8, 'pattern': r'((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))'}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'reference_trigger': {'key': 'referenceTrigger', 'type': 'TriggerReference'}, - 'offset': {'key': 'offset', 'type': 'str'}, - 'size': {'key': 'size', 'type': 'str'}, - } - - def __init__( - self, - *, - reference_trigger: "TriggerReference", - offset: Optional[str] = None, - size: Optional[str] = None, - **kwargs - ): - super(TumblingWindowTriggerDependencyReference, self).__init__(reference_trigger=reference_trigger, **kwargs) - self.type = 'TumblingWindowTriggerDependencyReference' # type: str - self.offset = offset - self.size = size - - -class TypeConversionSettings(msrest.serialization.Model): - """Type conversion settings. - - :param allow_data_truncation: Whether to allow data truncation when converting the data. Type: - boolean (or Expression with resultType boolean). - :type allow_data_truncation: any - :param treat_boolean_as_number: Whether to treat boolean values as numbers. Type: boolean (or - Expression with resultType boolean). - :type treat_boolean_as_number: any - :param date_time_format: The format for DateTime values. Type: string (or Expression with - resultType string). - :type date_time_format: any - :param date_time_offset_format: The format for DateTimeOffset values. Type: string (or - Expression with resultType string). - :type date_time_offset_format: any - :param time_span_format: The format for TimeSpan values. Type: string (or Expression with - resultType string). - :type time_span_format: any - :param culture: The culture used to convert data from/to string. Type: string (or Expression - with resultType string). - :type culture: any - """ - - _attribute_map = { - 'allow_data_truncation': {'key': 'allowDataTruncation', 'type': 'object'}, - 'treat_boolean_as_number': {'key': 'treatBooleanAsNumber', 'type': 'object'}, - 'date_time_format': {'key': 'dateTimeFormat', 'type': 'object'}, - 'date_time_offset_format': {'key': 'dateTimeOffsetFormat', 'type': 'object'}, - 'time_span_format': {'key': 'timeSpanFormat', 'type': 'object'}, - 'culture': {'key': 'culture', 'type': 'object'}, - } - - def __init__( - self, - *, - allow_data_truncation: Optional[Any] = None, - treat_boolean_as_number: Optional[Any] = None, - date_time_format: Optional[Any] = None, - date_time_offset_format: Optional[Any] = None, - time_span_format: Optional[Any] = None, - culture: Optional[Any] = None, - **kwargs - ): - super(TypeConversionSettings, self).__init__(**kwargs) - self.allow_data_truncation = allow_data_truncation - self.treat_boolean_as_number = treat_boolean_as_number - self.date_time_format = date_time_format - self.date_time_offset_format = date_time_offset_format - self.time_span_format = time_span_format - self.culture = culture - - -class UntilActivity(ControlActivity): - """This activity executes inner activities until the specified boolean expression results to true or timeout is reached, whichever is earlier. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param expression: Required. An expression that would evaluate to Boolean. The loop will - continue until this expression evaluates to true. - :type expression: ~azure.synapse.artifacts.v2020_12_01.models.Expression - :param timeout: Specifies the timeout for the activity to run. If there is no value specified, - it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or - Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). Type: string (or Expression with - resultType string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type timeout: any - :param activities: Required. List of activities to execute. - :type activities: list[~azure.synapse.artifacts.v2020_12_01.models.Activity] - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'expression': {'required': True}, - 'activities': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'expression': {'key': 'typeProperties.expression', 'type': 'Expression'}, - 'timeout': {'key': 'typeProperties.timeout', 'type': 'object'}, - 'activities': {'key': 'typeProperties.activities', 'type': '[Activity]'}, - } - - def __init__( - self, - *, - name: str, - expression: "Expression", - activities: List["Activity"], - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - timeout: Optional[Any] = None, - **kwargs - ): - super(UntilActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) - self.type = 'Until' # type: str - self.expression = expression - self.timeout = timeout - self.activities = activities - - -class UserProperty(msrest.serialization.Model): - """User property. - - All required parameters must be populated in order to send to Azure. - - :param name: Required. User property name. - :type name: str - :param value: Required. User property value. Type: string (or Expression with resultType - string). - :type value: any - """ - - _validation = { - 'name': {'required': True}, - 'value': {'required': True}, - } - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'value': {'key': 'value', 'type': 'object'}, - } - - def __init__( - self, - *, - name: str, - value: Any, - **kwargs - ): - super(UserProperty, self).__init__(**kwargs) - self.name = name - self.value = value - - -class ValidationActivity(ControlActivity): - """This activity verifies that an external resource exists. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param timeout: Specifies the timeout for the activity to run. If there is no value specified, - it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string (or - Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type timeout: any - :param sleep: A delay in seconds between validation attempts. If no value is specified, 10 - seconds will be used as the default. Type: integer (or Expression with resultType integer). - :type sleep: any - :param minimum_size: Can be used if dataset points to a file. The file must be greater than or - equal in size to the value specified. Type: integer (or Expression with resultType integer). - :type minimum_size: any - :param child_items: Can be used if dataset points to a folder. If set to true, the folder must - have at least one file. If set to false, the folder must be empty. Type: boolean (or Expression - with resultType boolean). - :type child_items: any - :param dataset: Required. Validation activity dataset reference. - :type dataset: ~azure.synapse.artifacts.v2020_12_01.models.DatasetReference - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'dataset': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'timeout': {'key': 'typeProperties.timeout', 'type': 'object'}, - 'sleep': {'key': 'typeProperties.sleep', 'type': 'object'}, - 'minimum_size': {'key': 'typeProperties.minimumSize', 'type': 'object'}, - 'child_items': {'key': 'typeProperties.childItems', 'type': 'object'}, - 'dataset': {'key': 'typeProperties.dataset', 'type': 'DatasetReference'}, - } - - def __init__( - self, - *, - name: str, - dataset: "DatasetReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - timeout: Optional[Any] = None, - sleep: Optional[Any] = None, - minimum_size: Optional[Any] = None, - child_items: Optional[Any] = None, - **kwargs - ): - super(ValidationActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) - self.type = 'Validation' # type: str - self.timeout = timeout - self.sleep = sleep - self.minimum_size = minimum_size - self.child_items = child_items - self.dataset = dataset - - -class VariableSpecification(msrest.serialization.Model): - """Definition of a single variable for a Pipeline. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Variable type. Possible values include: "String", "Bool", "Boolean", - "Array". - :type type: str or ~azure.synapse.artifacts.v2020_12_01.models.VariableType - :param default_value: Default value of variable. - :type default_value: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'default_value': {'key': 'defaultValue', 'type': 'object'}, - } - - def __init__( - self, - *, - type: Union[str, "VariableType"], - default_value: Optional[Any] = None, - **kwargs - ): - super(VariableSpecification, self).__init__(**kwargs) - self.type = type - self.default_value = default_value - - -class VerticaLinkedService(LinkedService): - """Vertica linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_string: An ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :type connection_string: any - :param pwd: The Azure key vault secret reference of password in connection string. - :type pwd: ~azure.synapse.artifacts.v2020_12_01.models.AzureKeyVaultSecretReference - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, - 'pwd': {'key': 'typeProperties.pwd', 'type': 'AzureKeyVaultSecretReference'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_string: Optional[Any] = None, - pwd: Optional["AzureKeyVaultSecretReference"] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(VerticaLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Vertica' # type: str - self.connection_string = connection_string - self.pwd = pwd - self.encrypted_credential = encrypted_credential - - -class VerticaSource(TabularSource): - """A copy activity Vertica source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(VerticaSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'VerticaSource' # type: str - self.query = query - - -class VerticaTableDataset(Dataset): - """Vertica dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: This property will be retired. Please consider using schema + table - properties instead. - :type table_name: any - :param table: The table name of the Vertica. Type: string (or Expression with resultType - string). - :type table: any - :param schema_type_properties_schema: The schema name of the Vertica. Type: string (or - Expression with resultType string). - :type schema_type_properties_schema: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - 'schema_type_properties_schema': {'key': 'typeProperties.schema', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - table: Optional[Any] = None, - schema_type_properties_schema: Optional[Any] = None, - **kwargs - ): - super(VerticaTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'VerticaTable' # type: str - self.table_name = table_name - self.table = table - self.schema_type_properties_schema = schema_type_properties_schema - - -class VirtualNetworkProfile(msrest.serialization.Model): - """Virtual Network Profile. - - :param compute_subnet_id: Subnet ID used for computes in workspace. - :type compute_subnet_id: str - """ - - _attribute_map = { - 'compute_subnet_id': {'key': 'computeSubnetId', 'type': 'str'}, - } - - def __init__( - self, - *, - compute_subnet_id: Optional[str] = None, - **kwargs - ): - super(VirtualNetworkProfile, self).__init__(**kwargs) - self.compute_subnet_id = compute_subnet_id - - -class WaitActivity(ControlActivity): - """This activity suspends pipeline execution for the specified interval. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param wait_time_in_seconds: Required. Duration in seconds. - :type wait_time_in_seconds: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'wait_time_in_seconds': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'wait_time_in_seconds': {'key': 'typeProperties.waitTimeInSeconds', 'type': 'object'}, - } - - def __init__( - self, - *, - name: str, - wait_time_in_seconds: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - **kwargs - ): - super(WaitActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) - self.type = 'Wait' # type: str - self.wait_time_in_seconds = wait_time_in_seconds - - -class WebActivity(ExecutionActivity): - """Web activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param linked_service_name: Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param policy: Activity policy. - :type policy: ~azure.synapse.artifacts.v2020_12_01.models.ActivityPolicy - :param method: Required. Rest API method for target endpoint. Possible values include: "GET", - "POST", "PUT", "DELETE". - :type method: str or ~azure.synapse.artifacts.v2020_12_01.models.WebActivityMethod - :param url: Required. Web activity target endpoint and path. Type: string (or Expression with - resultType string). - :type url: any - :param headers: Represents the headers that will be sent to the request. For example, to set - the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": - "application/json" }. Type: string (or Expression with resultType string). - :type headers: any - :param body: Represents the payload that will be sent to the endpoint. Required for POST/PUT - method, not allowed for GET method Type: string (or Expression with resultType string). - :type body: any - :param authentication: Authentication method used for calling the endpoint. - :type authentication: ~azure.synapse.artifacts.v2020_12_01.models.WebActivityAuthentication - :param datasets: List of datasets passed to web endpoint. - :type datasets: list[~azure.synapse.artifacts.v2020_12_01.models.DatasetReference] - :param linked_services: List of linked services passed to web endpoint. - :type linked_services: list[~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference] - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'method': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, - 'method': {'key': 'typeProperties.method', 'type': 'str'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'headers': {'key': 'typeProperties.headers', 'type': 'object'}, - 'body': {'key': 'typeProperties.body', 'type': 'object'}, - 'authentication': {'key': 'typeProperties.authentication', 'type': 'WebActivityAuthentication'}, - 'datasets': {'key': 'typeProperties.datasets', 'type': '[DatasetReference]'}, - 'linked_services': {'key': 'typeProperties.linkedServices', 'type': '[LinkedServiceReference]'}, - 'connect_via': {'key': 'typeProperties.connectVia', 'type': 'IntegrationRuntimeReference'}, - } - - def __init__( - self, - *, - name: str, - method: Union[str, "WebActivityMethod"], - url: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - linked_service_name: Optional["LinkedServiceReference"] = None, - policy: Optional["ActivityPolicy"] = None, - headers: Optional[Any] = None, - body: Optional[Any] = None, - authentication: Optional["WebActivityAuthentication"] = None, - datasets: Optional[List["DatasetReference"]] = None, - linked_services: Optional[List["LinkedServiceReference"]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - **kwargs - ): - super(WebActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) - self.type = 'WebActivity' # type: str - self.method = method - self.url = url - self.headers = headers - self.body = body - self.authentication = authentication - self.datasets = datasets - self.linked_services = linked_services - self.connect_via = connect_via - - -class WebActivityAuthentication(msrest.serialization.Model): - """Web activity authentication properties. - - All required parameters must be populated in order to send to Azure. - - :param type: Required. Web activity authentication (Basic/ClientCertificate/MSI). - :type type: str - :param pfx: Base64-encoded contents of a PFX file. - :type pfx: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param username: Web activity authentication user name for basic authentication. - :type username: str - :param password: Password for the PFX file or basic authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param resource: Resource for which Azure Auth token will be requested when using MSI - Authentication. - :type resource: str - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'pfx': {'key': 'pfx', 'type': 'SecretBase'}, - 'username': {'key': 'username', 'type': 'str'}, - 'password': {'key': 'password', 'type': 'SecretBase'}, - 'resource': {'key': 'resource', 'type': 'str'}, - } - - def __init__( - self, - *, - type: str, - pfx: Optional["SecretBase"] = None, - username: Optional[str] = None, - password: Optional["SecretBase"] = None, - resource: Optional[str] = None, - **kwargs - ): - super(WebActivityAuthentication, self).__init__(**kwargs) - self.type = type - self.pfx = pfx - self.username = username - self.password = password - self.resource = resource - - -class WebLinkedServiceTypeProperties(msrest.serialization.Model): - """Base definition of WebLinkedServiceTypeProperties, this typeProperties is polymorphic based on authenticationType, so not flattened in SDK models. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: WebAnonymousAuthentication, WebBasicAuthentication, WebClientCertificateAuthentication. - - All required parameters must be populated in order to send to Azure. - - :param url: Required. The URL of the web service endpoint, e.g. http://www.microsoft.com . - Type: string (or Expression with resultType string). - :type url: any - :param authentication_type: Required. Type of authentication used to connect to the web table - source.Constant filled by server. Possible values include: "Basic", "Anonymous", - "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.WebAuthenticationType - """ - - _validation = { - 'url': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'url': {'key': 'url', 'type': 'object'}, - 'authentication_type': {'key': 'authenticationType', 'type': 'str'}, - } - - _subtype_map = { - 'authentication_type': {'Anonymous': 'WebAnonymousAuthentication', 'Basic': 'WebBasicAuthentication', 'ClientCertificate': 'WebClientCertificateAuthentication'} - } - - def __init__( - self, - *, - url: Any, - **kwargs - ): - super(WebLinkedServiceTypeProperties, self).__init__(**kwargs) - self.url = url - self.authentication_type = None # type: Optional[str] - - -class WebAnonymousAuthentication(WebLinkedServiceTypeProperties): - """A WebLinkedService that uses anonymous authentication to communicate with an HTTP endpoint. - - All required parameters must be populated in order to send to Azure. - - :param url: Required. The URL of the web service endpoint, e.g. http://www.microsoft.com . - Type: string (or Expression with resultType string). - :type url: any - :param authentication_type: Required. Type of authentication used to connect to the web table - source.Constant filled by server. Possible values include: "Basic", "Anonymous", - "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.WebAuthenticationType - """ - - _validation = { - 'url': {'required': True}, - 'authentication_type': {'required': True}, - } - - _attribute_map = { - 'url': {'key': 'url', 'type': 'object'}, - 'authentication_type': {'key': 'authenticationType', 'type': 'str'}, - } - - def __init__( - self, - *, - url: Any, - **kwargs - ): - super(WebAnonymousAuthentication, self).__init__(url=url, **kwargs) - self.authentication_type = 'Anonymous' # type: str - - -class WebBasicAuthentication(WebLinkedServiceTypeProperties): - """A WebLinkedService that uses basic authentication to communicate with an HTTP endpoint. - - All required parameters must be populated in order to send to Azure. - - :param url: Required. The URL of the web service endpoint, e.g. http://www.microsoft.com . - Type: string (or Expression with resultType string). - :type url: any - :param authentication_type: Required. Type of authentication used to connect to the web table - source.Constant filled by server. Possible values include: "Basic", "Anonymous", - "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.WebAuthenticationType - :param username: Required. User name for Basic authentication. Type: string (or Expression with - resultType string). - :type username: any - :param password: Required. The password for Basic authentication. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - """ - - _validation = { - 'url': {'required': True}, - 'authentication_type': {'required': True}, - 'username': {'required': True}, - 'password': {'required': True}, - } - - _attribute_map = { - 'url': {'key': 'url', 'type': 'object'}, - 'authentication_type': {'key': 'authenticationType', 'type': 'str'}, - 'username': {'key': 'username', 'type': 'object'}, - 'password': {'key': 'password', 'type': 'SecretBase'}, - } - - def __init__( - self, - *, - url: Any, - username: Any, - password: "SecretBase", - **kwargs - ): - super(WebBasicAuthentication, self).__init__(url=url, **kwargs) - self.authentication_type = 'Basic' # type: str - self.username = username - self.password = password - - -class WebClientCertificateAuthentication(WebLinkedServiceTypeProperties): - """A WebLinkedService that uses client certificate based authentication to communicate with an HTTP endpoint. This scheme follows mutual authentication; the server must also provide valid credentials to the client. - - All required parameters must be populated in order to send to Azure. - - :param url: Required. The URL of the web service endpoint, e.g. http://www.microsoft.com . - Type: string (or Expression with resultType string). - :type url: any - :param authentication_type: Required. Type of authentication used to connect to the web table - source.Constant filled by server. Possible values include: "Basic", "Anonymous", - "ClientCertificate". - :type authentication_type: str or - ~azure.synapse.artifacts.v2020_12_01.models.WebAuthenticationType - :param pfx: Required. Base64-encoded contents of a PFX file. - :type pfx: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param password: Required. Password for the PFX file. - :type password: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - """ - - _validation = { - 'url': {'required': True}, - 'authentication_type': {'required': True}, - 'pfx': {'required': True}, - 'password': {'required': True}, - } - - _attribute_map = { - 'url': {'key': 'url', 'type': 'object'}, - 'authentication_type': {'key': 'authenticationType', 'type': 'str'}, - 'pfx': {'key': 'pfx', 'type': 'SecretBase'}, - 'password': {'key': 'password', 'type': 'SecretBase'}, - } - - def __init__( - self, - *, - url: Any, - pfx: "SecretBase", - password: "SecretBase", - **kwargs - ): - super(WebClientCertificateAuthentication, self).__init__(url=url, **kwargs) - self.authentication_type = 'ClientCertificate' # type: str - self.pfx = pfx - self.password = password - - -class WebHookActivity(ControlActivity): - """WebHook activity. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param name: Required. Activity name. - :type name: str - :param type: Required. Type of activity.Constant filled by server. - :type type: str - :param description: Activity description. - :type description: str - :param depends_on: Activity depends on condition. - :type depends_on: list[~azure.synapse.artifacts.v2020_12_01.models.ActivityDependency] - :param user_properties: Activity user properties. - :type user_properties: list[~azure.synapse.artifacts.v2020_12_01.models.UserProperty] - :param method: Required. Rest API method for target endpoint. Possible values include: "POST". - :type method: str or ~azure.synapse.artifacts.v2020_12_01.models.WebHookActivityMethod - :param url: Required. WebHook activity target endpoint and path. Type: string (or Expression - with resultType string). - :type url: any - :param timeout: The timeout within which the webhook should be called back. If there is no - value specified, it defaults to 10 minutes. Type: string. Pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type timeout: str - :param headers: Represents the headers that will be sent to the request. For example, to set - the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": - "application/json" }. Type: string (or Expression with resultType string). - :type headers: any - :param body: Represents the payload that will be sent to the endpoint. Required for POST/PUT - method, not allowed for GET method Type: string (or Expression with resultType string). - :type body: any - :param authentication: Authentication method used for calling the endpoint. - :type authentication: ~azure.synapse.artifacts.v2020_12_01.models.WebActivityAuthentication - :param report_status_on_call_back: When set to true, statusCode, output and error in callback - request body will be consumed by activity. The activity can be marked as failed by setting - statusCode >= 400 in callback request. Default is false. Type: boolean (or Expression with - resultType boolean). - :type report_status_on_call_back: any - """ - - _validation = { - 'name': {'required': True}, - 'type': {'required': True}, - 'method': {'required': True}, - 'url': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, - 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, - 'method': {'key': 'typeProperties.method', 'type': 'str'}, - 'url': {'key': 'typeProperties.url', 'type': 'object'}, - 'timeout': {'key': 'typeProperties.timeout', 'type': 'str'}, - 'headers': {'key': 'typeProperties.headers', 'type': 'object'}, - 'body': {'key': 'typeProperties.body', 'type': 'object'}, - 'authentication': {'key': 'typeProperties.authentication', 'type': 'WebActivityAuthentication'}, - 'report_status_on_call_back': {'key': 'typeProperties.reportStatusOnCallBack', 'type': 'object'}, - } - - def __init__( - self, - *, - name: str, - method: Union[str, "WebHookActivityMethod"], - url: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - depends_on: Optional[List["ActivityDependency"]] = None, - user_properties: Optional[List["UserProperty"]] = None, - timeout: Optional[str] = None, - headers: Optional[Any] = None, - body: Optional[Any] = None, - authentication: Optional["WebActivityAuthentication"] = None, - report_status_on_call_back: Optional[Any] = None, - **kwargs - ): - super(WebHookActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) - self.type = 'WebHook' # type: str - self.method = method - self.url = url - self.timeout = timeout - self.headers = headers - self.body = body - self.authentication = authentication - self.report_status_on_call_back = report_status_on_call_back - - -class WebLinkedService(LinkedService): - """Web linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param type_properties: Required. Web linked service properties. - :type type_properties: - ~azure.synapse.artifacts.v2020_12_01.models.WebLinkedServiceTypeProperties - """ - - _validation = { - 'type': {'required': True}, - 'type_properties': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'type_properties': {'key': 'typeProperties', 'type': 'WebLinkedServiceTypeProperties'}, - } - - def __init__( - self, - *, - type_properties: "WebLinkedServiceTypeProperties", - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - **kwargs - ): - super(WebLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Web' # type: str - self.type_properties = type_properties - - -class WebSource(CopySource): - """A copy activity source for web page table. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(WebSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'WebSource' # type: str - self.additional_columns = additional_columns - - -class WebTableDataset(Dataset): - """The dataset points to a HTML table in the web page. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param index: Required. The zero-based index of the table in the web page. Type: integer (or - Expression with resultType integer), minimum: 0. - :type index: any - :param path: The relative URL to the web page from the linked service URL. Type: string (or - Expression with resultType string). - :type path: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - 'index': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'index': {'key': 'typeProperties.index', 'type': 'object'}, - 'path': {'key': 'typeProperties.path', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - index: Any, - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - path: Optional[Any] = None, - **kwargs - ): - super(WebTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'WebTable' # type: str - self.index = index - self.path = path - - -class Workspace(TrackedResource): - """A workspace. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar id: Fully qualified resource ID for the resource. Ex - - /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. - :vartype id: str - :ivar name: The name of the resource. - :vartype name: str - :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or - "Microsoft.Storage/storageAccounts". - :vartype type: str - :param tags: A set of tags. Resource tags. - :type tags: dict[str, str] - :param location: Required. The geo-location where the resource lives. - :type location: str - :param identity: Identity of the workspace. - :type identity: ~azure.synapse.artifacts.v2020_12_01.models.ManagedIdentity - :param default_data_lake_storage: Workspace default data lake storage account details. - :type default_data_lake_storage: - ~azure.synapse.artifacts.v2020_12_01.models.DataLakeStorageAccountDetails - :param sql_administrator_login_password: SQL administrator login password. - :type sql_administrator_login_password: str - :param managed_resource_group_name: Workspace managed resource group. The resource group name - uniquely identifies the resource group within the user subscriptionId. The resource group name - must be no longer than 90 characters long, and must be alphanumeric characters - (Char.IsLetterOrDigit()) and '-', '_', '(', ')' and'.'. Note that the name cannot end with '.'. - :type managed_resource_group_name: str - :ivar provisioning_state: Resource provisioning state. - :vartype provisioning_state: str - :param sql_administrator_login: Login for workspace SQL active directory administrator. - :type sql_administrator_login: str - :param virtual_network_profile: Virtual Network profile. - :type virtual_network_profile: - ~azure.synapse.artifacts.v2020_12_01.models.VirtualNetworkProfile - :param connectivity_endpoints: Connectivity endpoints. - :type connectivity_endpoints: dict[str, str] - :param managed_virtual_network: Setting this to 'default' will ensure that all compute for this - workspace is in a virtual network managed on behalf of the user. - :type managed_virtual_network: str - :param private_endpoint_connections: Private endpoint connections to the workspace. - :type private_endpoint_connections: - list[~azure.synapse.artifacts.v2020_12_01.models.PrivateEndpointConnection] - :param encryption: The encryption details of the workspace. - :type encryption: ~azure.synapse.artifacts.v2020_12_01.models.EncryptionDetails - :ivar workspace_uid: The workspace unique identifier. - :vartype workspace_uid: str - :ivar extra_properties: Workspace level configs and feature flags. - :vartype extra_properties: dict[str, any] - :param managed_virtual_network_settings: Managed Virtual Network Settings. - :type managed_virtual_network_settings: - ~azure.synapse.artifacts.v2020_12_01.models.ManagedVirtualNetworkSettings - :param workspace_repository_configuration: Git integration settings. - :type workspace_repository_configuration: - ~azure.synapse.artifacts.v2020_12_01.models.WorkspaceRepositoryConfiguration - :param purview_configuration: Purview Configuration. - :type purview_configuration: ~azure.synapse.artifacts.v2020_12_01.models.PurviewConfiguration - :ivar adla_resource_id: The ADLA resource ID. - :vartype adla_resource_id: str - """ - - _validation = { - 'id': {'readonly': True}, - 'name': {'readonly': True}, - 'type': {'readonly': True}, - 'location': {'required': True}, - 'provisioning_state': {'readonly': True}, - 'workspace_uid': {'readonly': True}, - 'extra_properties': {'readonly': True}, - 'adla_resource_id': {'readonly': True}, - } - - _attribute_map = { - 'id': {'key': 'id', 'type': 'str'}, - 'name': {'key': 'name', 'type': 'str'}, - 'type': {'key': 'type', 'type': 'str'}, - 'tags': {'key': 'tags', 'type': '{str}'}, - 'location': {'key': 'location', 'type': 'str'}, - 'identity': {'key': 'identity', 'type': 'ManagedIdentity'}, - 'default_data_lake_storage': {'key': 'properties.defaultDataLakeStorage', 'type': 'DataLakeStorageAccountDetails'}, - 'sql_administrator_login_password': {'key': 'properties.sqlAdministratorLoginPassword', 'type': 'str'}, - 'managed_resource_group_name': {'key': 'properties.managedResourceGroupName', 'type': 'str'}, - 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, - 'sql_administrator_login': {'key': 'properties.sqlAdministratorLogin', 'type': 'str'}, - 'virtual_network_profile': {'key': 'properties.virtualNetworkProfile', 'type': 'VirtualNetworkProfile'}, - 'connectivity_endpoints': {'key': 'properties.connectivityEndpoints', 'type': '{str}'}, - 'managed_virtual_network': {'key': 'properties.managedVirtualNetwork', 'type': 'str'}, - 'private_endpoint_connections': {'key': 'properties.privateEndpointConnections', 'type': '[PrivateEndpointConnection]'}, - 'encryption': {'key': 'properties.encryption', 'type': 'EncryptionDetails'}, - 'workspace_uid': {'key': 'properties.workspaceUID', 'type': 'str'}, - 'extra_properties': {'key': 'properties.extraProperties', 'type': '{object}'}, - 'managed_virtual_network_settings': {'key': 'properties.managedVirtualNetworkSettings', 'type': 'ManagedVirtualNetworkSettings'}, - 'workspace_repository_configuration': {'key': 'properties.workspaceRepositoryConfiguration', 'type': 'WorkspaceRepositoryConfiguration'}, - 'purview_configuration': {'key': 'properties.purviewConfiguration', 'type': 'PurviewConfiguration'}, - 'adla_resource_id': {'key': 'properties.adlaResourceId', 'type': 'str'}, - } - - def __init__( - self, - *, - location: str, - tags: Optional[Dict[str, str]] = None, - identity: Optional["ManagedIdentity"] = None, - default_data_lake_storage: Optional["DataLakeStorageAccountDetails"] = None, - sql_administrator_login_password: Optional[str] = None, - managed_resource_group_name: Optional[str] = None, - sql_administrator_login: Optional[str] = None, - virtual_network_profile: Optional["VirtualNetworkProfile"] = None, - connectivity_endpoints: Optional[Dict[str, str]] = None, - managed_virtual_network: Optional[str] = None, - private_endpoint_connections: Optional[List["PrivateEndpointConnection"]] = None, - encryption: Optional["EncryptionDetails"] = None, - managed_virtual_network_settings: Optional["ManagedVirtualNetworkSettings"] = None, - workspace_repository_configuration: Optional["WorkspaceRepositoryConfiguration"] = None, - purview_configuration: Optional["PurviewConfiguration"] = None, - **kwargs - ): - super(Workspace, self).__init__(tags=tags, location=location, **kwargs) - self.identity = identity - self.default_data_lake_storage = default_data_lake_storage - self.sql_administrator_login_password = sql_administrator_login_password - self.managed_resource_group_name = managed_resource_group_name - self.provisioning_state = None - self.sql_administrator_login = sql_administrator_login - self.virtual_network_profile = virtual_network_profile - self.connectivity_endpoints = connectivity_endpoints - self.managed_virtual_network = managed_virtual_network - self.private_endpoint_connections = private_endpoint_connections - self.encryption = encryption - self.workspace_uid = None - self.extra_properties = None - self.managed_virtual_network_settings = managed_virtual_network_settings - self.workspace_repository_configuration = workspace_repository_configuration - self.purview_configuration = purview_configuration - self.adla_resource_id = None - - -class WorkspaceIdentity(msrest.serialization.Model): - """Identity properties of the workspace resource. - - Variables are only populated by the server, and will be ignored when sending a request. - - All required parameters must be populated in order to send to Azure. - - :ivar type: The identity type. Currently the only supported type is 'SystemAssigned'. Has - constant value: "SystemAssigned". - :vartype type: str - :ivar principal_id: The principal id of the identity. - :vartype principal_id: str - :ivar tenant_id: The client tenant id of the identity. - :vartype tenant_id: str - """ - - _validation = { - 'type': {'required': True, 'constant': True}, - 'principal_id': {'readonly': True}, - 'tenant_id': {'readonly': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'principal_id': {'key': 'principalId', 'type': 'str'}, - 'tenant_id': {'key': 'tenantId', 'type': 'str'}, - } - - type = "SystemAssigned" - - def __init__( - self, - **kwargs - ): - super(WorkspaceIdentity, self).__init__(**kwargs) - self.principal_id = None - self.tenant_id = None - - -class WorkspaceKeyDetails(msrest.serialization.Model): - """Details of the customer managed key associated with the workspace. - - :param name: Workspace Key sub-resource name. - :type name: str - :param key_vault_url: Workspace Key sub-resource key vault url. - :type key_vault_url: str - """ - - _attribute_map = { - 'name': {'key': 'name', 'type': 'str'}, - 'key_vault_url': {'key': 'keyVaultUrl', 'type': 'str'}, - } - - def __init__( - self, - *, - name: Optional[str] = None, - key_vault_url: Optional[str] = None, - **kwargs - ): - super(WorkspaceKeyDetails, self).__init__(**kwargs) - self.name = name - self.key_vault_url = key_vault_url - - -class WorkspaceRepositoryConfiguration(msrest.serialization.Model): - """Git integration settings. - - :param type: Type of workspace repositoryID configuration. Example WorkspaceVSTSConfiguration, - WorkspaceGitHubConfiguration. - :type type: str - :param host_name: GitHub Enterprise host name. For example: https://github.mydomain.com. - :type host_name: str - :param account_name: Account name. - :type account_name: str - :param project_name: VSTS project name. - :type project_name: str - :param repository_name: Repository name. - :type repository_name: str - :param collaboration_branch: Collaboration branch. - :type collaboration_branch: str - :param root_folder: Root folder to use in the repository. - :type root_folder: str - :param last_commit_id: The last commit ID. - :type last_commit_id: str - :param tenant_id: The VSTS tenant ID. - :type tenant_id: str - :param client_id: GitHub bring your own app client id. - :type client_id: str - :param client_secret: GitHub bring your own app client secret information. - :type client_secret: ~azure.synapse.artifacts.v2020_12_01.models.GitHubClientSecret - """ - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'host_name': {'key': 'hostName', 'type': 'str'}, - 'account_name': {'key': 'accountName', 'type': 'str'}, - 'project_name': {'key': 'projectName', 'type': 'str'}, - 'repository_name': {'key': 'repositoryName', 'type': 'str'}, - 'collaboration_branch': {'key': 'collaborationBranch', 'type': 'str'}, - 'root_folder': {'key': 'rootFolder', 'type': 'str'}, - 'last_commit_id': {'key': 'lastCommitId', 'type': 'str'}, - 'tenant_id': {'key': 'tenantId', 'type': 'str'}, - 'client_id': {'key': 'clientId', 'type': 'str'}, - 'client_secret': {'key': 'clientSecret', 'type': 'GitHubClientSecret'}, - } - - def __init__( - self, - *, - type: Optional[str] = None, - host_name: Optional[str] = None, - account_name: Optional[str] = None, - project_name: Optional[str] = None, - repository_name: Optional[str] = None, - collaboration_branch: Optional[str] = None, - root_folder: Optional[str] = None, - last_commit_id: Optional[str] = None, - tenant_id: Optional[str] = None, - client_id: Optional[str] = None, - client_secret: Optional["GitHubClientSecret"] = None, - **kwargs - ): - super(WorkspaceRepositoryConfiguration, self).__init__(**kwargs) - self.type = type - self.host_name = host_name - self.account_name = account_name - self.project_name = project_name - self.repository_name = repository_name - self.collaboration_branch = collaboration_branch - self.root_folder = root_folder - self.last_commit_id = last_commit_id - self.tenant_id = tenant_id - self.client_id = client_id - self.client_secret = client_secret - - -class WorkspaceUpdateParameters(msrest.serialization.Model): - """Parameters for updating a workspace resource. - - :param tags: A set of tags. The resource tags. - :type tags: dict[str, str] - :param identity: Managed service identity of the workspace. - :type identity: ~azure.synapse.artifacts.v2020_12_01.models.WorkspaceIdentity - """ - - _attribute_map = { - 'tags': {'key': 'tags', 'type': '{str}'}, - 'identity': {'key': 'identity', 'type': 'WorkspaceIdentity'}, - } - - def __init__( - self, - *, - tags: Optional[Dict[str, str]] = None, - identity: Optional["WorkspaceIdentity"] = None, - **kwargs - ): - super(WorkspaceUpdateParameters, self).__init__(**kwargs) - self.tags = tags - self.identity = identity - - -class XeroLinkedService(LinkedService): - """Xero Service linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_properties: Properties used to connect to Xero. It is mutually exclusive with - any other properties in the linked service. Type: object. - :type connection_properties: any - :param host: Required. The endpoint of the Xero server. (i.e. api.xero.com). - :type host: any - :param consumer_key: The consumer key associated with the Xero application. - :type consumer_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param private_key: The private key from the .pem file that was generated for your Xero private - application. You must include all the text from the .pem file, including the Unix line endings( - ). - :type private_key: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'host': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_properties': {'key': 'typeProperties.connectionProperties', 'type': 'object'}, - 'host': {'key': 'typeProperties.host', 'type': 'object'}, - 'consumer_key': {'key': 'typeProperties.consumerKey', 'type': 'SecretBase'}, - 'private_key': {'key': 'typeProperties.privateKey', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - host: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_properties: Optional[Any] = None, - consumer_key: Optional["SecretBase"] = None, - private_key: Optional["SecretBase"] = None, - use_encrypted_endpoints: Optional[Any] = None, - use_host_verification: Optional[Any] = None, - use_peer_verification: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(XeroLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Xero' # type: str - self.connection_properties = connection_properties - self.host = host - self.consumer_key = consumer_key - self.private_key = private_key - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential - - -class XeroObjectDataset(Dataset): - """Xero Service dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(XeroObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'XeroObject' # type: str - self.table_name = table_name - - -class XeroSource(TabularSource): - """A copy activity Xero Service source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(XeroSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'XeroSource' # type: str - self.query = query - - -class XmlDataset(Dataset): - """Xml dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param location: The location of the json data storage. - :type location: ~azure.synapse.artifacts.v2020_12_01.models.DatasetLocation - :param encoding_name: The code page name of the preferred encoding. If not specified, the - default value is UTF-8, unless BOM denotes another Unicode encoding. Refer to the name column - of the table in the following link to set supported values: - https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string (or Expression with - resultType string). - :type encoding_name: any - :param null_value: The null value string. Type: string (or Expression with resultType string). - :type null_value: any - :param compression: The data compression method used for the json dataset. - :type compression: ~azure.synapse.artifacts.v2020_12_01.models.DatasetCompression - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'location': {'key': 'typeProperties.location', 'type': 'DatasetLocation'}, - 'encoding_name': {'key': 'typeProperties.encodingName', 'type': 'object'}, - 'null_value': {'key': 'typeProperties.nullValue', 'type': 'object'}, - 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - location: Optional["DatasetLocation"] = None, - encoding_name: Optional[Any] = None, - null_value: Optional[Any] = None, - compression: Optional["DatasetCompression"] = None, - **kwargs - ): - super(XmlDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'Xml' # type: str - self.location = location - self.encoding_name = encoding_name - self.null_value = null_value - self.compression = compression - - -class XmlReadSettings(FormatReadSettings): - """Xml read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The read setting type.Constant filled by server. - :type type: str - :param compression_properties: Compression settings. - :type compression_properties: - ~azure.synapse.artifacts.v2020_12_01.models.CompressionReadSettings - :param validation_mode: Indicates what validation method is used when reading the xml files. - Allowed values: 'none', 'xsd', or 'dtd'. Type: string (or Expression with resultType string). - :type validation_mode: any - :param detect_data_type: Indicates whether type detection is enabled when reading the xml - files. Type: boolean (or Expression with resultType boolean). - :type detect_data_type: any - :param namespaces: Indicates whether namespace is enabled when reading the xml files. Type: - boolean (or Expression with resultType boolean). - :type namespaces: any - :param namespace_prefixes: Namespace uri to prefix mappings to override the prefixes in column - names when namespace is enabled, if no prefix is defined for a namespace uri, the prefix of xml - element/attribute name in the xml data file will be used. Example: - "{"http://www.example.com/xml":"prefix"}" Type: object (or Expression with resultType object). - :type namespace_prefixes: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'compression_properties': {'key': 'compressionProperties', 'type': 'CompressionReadSettings'}, - 'validation_mode': {'key': 'validationMode', 'type': 'object'}, - 'detect_data_type': {'key': 'detectDataType', 'type': 'object'}, - 'namespaces': {'key': 'namespaces', 'type': 'object'}, - 'namespace_prefixes': {'key': 'namespacePrefixes', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - compression_properties: Optional["CompressionReadSettings"] = None, - validation_mode: Optional[Any] = None, - detect_data_type: Optional[Any] = None, - namespaces: Optional[Any] = None, - namespace_prefixes: Optional[Any] = None, - **kwargs - ): - super(XmlReadSettings, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'XmlReadSettings' # type: str - self.compression_properties = compression_properties - self.validation_mode = validation_mode - self.detect_data_type = detect_data_type - self.namespaces = namespaces - self.namespace_prefixes = namespace_prefixes - - -class XmlSource(CopySource): - """A copy activity Xml source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param store_settings: Xml store settings. - :type store_settings: ~azure.synapse.artifacts.v2020_12_01.models.StoreReadSettings - :param format_settings: Xml format settings. - :type format_settings: ~azure.synapse.artifacts.v2020_12_01.models.XmlReadSettings - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'format_settings': {'key': 'formatSettings', 'type': 'XmlReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - store_settings: Optional["StoreReadSettings"] = None, - format_settings: Optional["XmlReadSettings"] = None, - additional_columns: Optional[Any] = None, - **kwargs - ): - super(XmlSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.type = 'XmlSource' # type: str - self.store_settings = store_settings - self.format_settings = format_settings - self.additional_columns = additional_columns - - -class ZipDeflateReadSettings(CompressionReadSettings): - """The ZipDeflate compression read settings. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. The Compression setting type.Constant filled by server. - :type type: str - :param preserve_zip_file_name_as_folder: Preserve the zip file name as folder path. Type: - boolean (or Expression with resultType boolean). - :type preserve_zip_file_name_as_folder: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'preserve_zip_file_name_as_folder': {'key': 'preserveZipFileNameAsFolder', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - preserve_zip_file_name_as_folder: Optional[Any] = None, - **kwargs - ): - super(ZipDeflateReadSettings, self).__init__(additional_properties=additional_properties, **kwargs) - self.type = 'ZipDeflateReadSettings' # type: str - self.preserve_zip_file_name_as_folder = preserve_zip_file_name_as_folder - - -class ZohoLinkedService(LinkedService): - """Zoho server linked service. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of linked service.Constant filled by server. - :type type: str - :param connect_via: The integration runtime reference. - :type connect_via: ~azure.synapse.artifacts.v2020_12_01.models.IntegrationRuntimeReference - :param description: Linked service description. - :type description: str - :param parameters: Parameters for linked service. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the linked service. - :type annotations: list[any] - :param connection_properties: Properties used to connect to Zoho. It is mutually exclusive with - any other properties in the linked service. Type: object. - :type connection_properties: any - :param endpoint: Required. The endpoint of the Zoho server. (i.e. crm.zoho.com/crm/private). - :type endpoint: any - :param access_token: The access token for Zoho authentication. - :type access_token: ~azure.synapse.artifacts.v2020_12_01.models.SecretBase - :param use_encrypted_endpoints: Specifies whether the data source endpoints are encrypted using - HTTPS. The default value is true. - :type use_encrypted_endpoints: any - :param use_host_verification: Specifies whether to require the host name in the server's - certificate to match the host name of the server when connecting over SSL. The default value is - true. - :type use_host_verification: any - :param use_peer_verification: Specifies whether to verify the identity of the server when - connecting over SSL. The default value is true. - :type use_peer_verification: any - :param encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string (or Expression with - resultType string). - :type encrypted_credential: any - """ - - _validation = { - 'type': {'required': True}, - 'endpoint': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, - 'description': {'key': 'description', 'type': 'str'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'connection_properties': {'key': 'typeProperties.connectionProperties', 'type': 'object'}, - 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, - 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, - 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, - 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, - 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, - 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, - } - - def __init__( - self, - *, - endpoint: Any, - additional_properties: Optional[Dict[str, Any]] = None, - connect_via: Optional["IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - connection_properties: Optional[Any] = None, - access_token: Optional["SecretBase"] = None, - use_encrypted_endpoints: Optional[Any] = None, - use_host_verification: Optional[Any] = None, - use_peer_verification: Optional[Any] = None, - encrypted_credential: Optional[Any] = None, - **kwargs - ): - super(ZohoLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.type = 'Zoho' # type: str - self.connection_properties = connection_properties - self.endpoint = endpoint - self.access_token = access_token - self.use_encrypted_endpoints = use_encrypted_endpoints - self.use_host_verification = use_host_verification - self.use_peer_verification = use_peer_verification - self.encrypted_credential = encrypted_credential - - -class ZohoObjectDataset(Dataset): - """Zoho server dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Type of dataset.Constant filled by server. - :type type: str - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :type structure: any - :param schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :type schema: any - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: ~azure.synapse.artifacts.v2020_12_01.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, ~azure.synapse.artifacts.v2020_12_01.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the Dataset. - :type annotations: list[any] - :param folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :type folder: ~azure.synapse.artifacts.v2020_12_01.models.DatasetFolder - :param table_name: The table name. Type: string (or Expression with resultType string). - :type table_name: any - """ - - _validation = { - 'type': {'required': True}, - 'linked_service_name': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, - } - - def __init__( - self, - *, - linked_service_name: "LinkedServiceReference", - additional_properties: Optional[Dict[str, Any]] = None, - description: Optional[str] = None, - structure: Optional[Any] = None, - schema: Optional[Any] = None, - parameters: Optional[Dict[str, "ParameterSpecification"]] = None, - annotations: Optional[List[Any]] = None, - folder: Optional["DatasetFolder"] = None, - table_name: Optional[Any] = None, - **kwargs - ): - super(ZohoObjectDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.type = 'ZohoObject' # type: str - self.table_name = table_name - - -class ZohoSource(TabularSource): - """A copy activity Zoho server source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are deserialized to this - collection. - :type additional_properties: dict[str, any] - :param type: Required. Copy source type.Constant filled by server. - :type type: str - :param source_retry_count: Source retry count. Type: integer (or Expression with resultType - integer). - :type source_retry_count: any - :param source_retry_wait: Source retry wait. Type: string (or Expression with resultType - string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: any - :param max_concurrent_connections: The maximum concurrent connection count for the source data - store. Type: integer (or Expression with resultType integer). - :type max_concurrent_connections: any - :param query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type query_timeout: any - :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). - :type additional_columns: any - :param query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :type query: any - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, - 'query': {'key': 'query', 'type': 'object'}, - } - - def __init__( - self, - *, - additional_properties: Optional[Dict[str, Any]] = None, - source_retry_count: Optional[Any] = None, - source_retry_wait: Optional[Any] = None, - max_concurrent_connections: Optional[Any] = None, - query_timeout: Optional[Any] = None, - additional_columns: Optional[Any] = None, - query: Optional[Any] = None, - **kwargs - ): - super(ZohoSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, query_timeout=query_timeout, additional_columns=additional_columns, **kwargs) - self.type = 'ZohoSource' # type: str - self.query = query diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/__init__.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/__init__.py deleted file mode 100644 index ddf9d04482ce..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/__init__.py +++ /dev/null @@ -1,51 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from ._linked_service_operations import LinkedServiceOperations -from ._dataset_operations import DatasetOperations -from ._pipeline_operations import PipelineOperations -from ._pipeline_run_operations import PipelineRunOperations -from ._trigger_operations import TriggerOperations -from ._trigger_run_operations import TriggerRunOperations -from ._data_flow_operations import DataFlowOperations -from ._data_flow_debug_session_operations import DataFlowDebugSessionOperations -from ._sql_script_operations import SqlScriptOperations -from ._spark_job_definition_operations import SparkJobDefinitionOperations -from ._notebook_operations import NotebookOperations -from ._notebook_operation_result_operations import NotebookOperationResultOperations -from ._big_data_pools_operations import BigDataPoolsOperations -from ._workspace_git_repo_management_operations import WorkspaceGitRepoManagementOperations -from ._integration_runtimes_operations import IntegrationRuntimesOperations -from ._library_operations import LibraryOperations -from ._operation_result_operations import OperationResultOperations -from ._operation_status_operations import OperationStatusOperations -from ._sql_pools_operations import SqlPoolsOperations -from ._workspace_operations import WorkspaceOperations - -__all__ = [ - 'LinkedServiceOperations', - 'DatasetOperations', - 'PipelineOperations', - 'PipelineRunOperations', - 'TriggerOperations', - 'TriggerRunOperations', - 'DataFlowOperations', - 'DataFlowDebugSessionOperations', - 'SqlScriptOperations', - 'SparkJobDefinitionOperations', - 'NotebookOperations', - 'NotebookOperationResultOperations', - 'BigDataPoolsOperations', - 'WorkspaceGitRepoManagementOperations', - 'IntegrationRuntimesOperations', - 'LibraryOperations', - 'OperationResultOperations', - 'OperationStatusOperations', - 'SqlPoolsOperations', - 'WorkspaceOperations', -] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_operation_result_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_operation_result_operations.py deleted file mode 100644 index fc1aff14f381..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_operation_result_operations.py +++ /dev/null @@ -1,97 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Optional, TypeVar - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class OperationResultOperations(object): - """OperationResultOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get( - self, - operation_id, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - """Get operation result. - - :param operation_id: Operation ID. - :type operation_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2020-12-01" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'operationId': self._serialize.url("operation_id", operation_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 201, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - get.metadata = {'url': '/operationResults/{operationId}'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_operation_status_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_operation_status_operations.py deleted file mode 100644 index 35206f8e1071..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/operations/_operation_status_operations.py +++ /dev/null @@ -1,97 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Optional, TypeVar - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class OperationStatusOperations(object): - """OperationStatusOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2020_12_01.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get( - self, - operation_id, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - """Get operation status. - - :param operation_id: Operation ID. - :type operation_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2020-12-01" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'operationId': self._serialize.url("operation_id", operation_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 201, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - get.metadata = {'url': '/operationStatuses/{operationId}'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/py.typed b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/py.typed deleted file mode 100644 index e5aff4f83af8..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2020_12_01/py.typed +++ /dev/null @@ -1 +0,0 @@ -# Marker file for PEP 561. \ No newline at end of file diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/__init__.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/__init__.py deleted file mode 100644 index 03703237ffc5..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/__init__.py +++ /dev/null @@ -1,19 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from ._artifacts_client import ArtifactsClient -from ._version import VERSION - -__version__ = VERSION -__all__ = ['ArtifactsClient'] - -try: - from ._patch import patch_sdk # type: ignore - patch_sdk() -except ImportError: - pass diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/_artifacts_client.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/_artifacts_client.py deleted file mode 100644 index 7c97299c7535..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/_artifacts_client.py +++ /dev/null @@ -1,186 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from typing import TYPE_CHECKING - -from azure.core import PipelineClient -from msrest import Deserializer, Serializer - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any - - from azure.core.credentials import TokenCredential - from azure.core.pipeline.transport import HttpRequest, HttpResponse - -from ._configuration import ArtifactsClientConfiguration -from .operations import LinkedServiceOperations -from .operations import DatasetOperations -from .operations import PipelineOperations -from .operations import PipelineRunOperations -from .operations import TriggerOperations -from .operations import TriggerRunOperations -from .operations import DataFlowOperations -from .operations import DataFlowDebugSessionOperations -from .operations import SqlScriptOperations -from .operations import SparkJobDefinitionOperations -from .operations import NotebookOperations -from .operations import NotebookOperationResultOperations -from .operations import SparkConfigurationOperations -from .operations import BigDataPoolsOperations -from .operations import WorkspaceGitRepoManagementOperations -from .operations import IntegrationRuntimesOperations -from .operations import LibraryOperations -from .operations import OperationResultOperations -from .operations import OperationStatusOperations -from .operations import SqlPoolsOperations -from .operations import WorkspaceOperations -from . import models - - -class ArtifactsClient(object): - """ArtifactsClient. - - :ivar linked_service: LinkedServiceOperations operations - :vartype linked_service: azure.synapse.artifacts.v2021_06_01_preview.operations.LinkedServiceOperations - :ivar dataset: DatasetOperations operations - :vartype dataset: azure.synapse.artifacts.v2021_06_01_preview.operations.DatasetOperations - :ivar pipeline: PipelineOperations operations - :vartype pipeline: azure.synapse.artifacts.v2021_06_01_preview.operations.PipelineOperations - :ivar pipeline_run: PipelineRunOperations operations - :vartype pipeline_run: azure.synapse.artifacts.v2021_06_01_preview.operations.PipelineRunOperations - :ivar trigger: TriggerOperations operations - :vartype trigger: azure.synapse.artifacts.v2021_06_01_preview.operations.TriggerOperations - :ivar trigger_run: TriggerRunOperations operations - :vartype trigger_run: azure.synapse.artifacts.v2021_06_01_preview.operations.TriggerRunOperations - :ivar data_flow: DataFlowOperations operations - :vartype data_flow: azure.synapse.artifacts.v2021_06_01_preview.operations.DataFlowOperations - :ivar data_flow_debug_session: DataFlowDebugSessionOperations operations - :vartype data_flow_debug_session: azure.synapse.artifacts.v2021_06_01_preview.operations.DataFlowDebugSessionOperations - :ivar sql_script: SqlScriptOperations operations - :vartype sql_script: azure.synapse.artifacts.v2021_06_01_preview.operations.SqlScriptOperations - :ivar spark_job_definition: SparkJobDefinitionOperations operations - :vartype spark_job_definition: azure.synapse.artifacts.v2021_06_01_preview.operations.SparkJobDefinitionOperations - :ivar notebook: NotebookOperations operations - :vartype notebook: azure.synapse.artifacts.v2021_06_01_preview.operations.NotebookOperations - :ivar notebook_operation_result: NotebookOperationResultOperations operations - :vartype notebook_operation_result: azure.synapse.artifacts.v2021_06_01_preview.operations.NotebookOperationResultOperations - :ivar spark_configuration: SparkConfigurationOperations operations - :vartype spark_configuration: azure.synapse.artifacts.v2021_06_01_preview.operations.SparkConfigurationOperations - :ivar big_data_pools: BigDataPoolsOperations operations - :vartype big_data_pools: azure.synapse.artifacts.v2021_06_01_preview.operations.BigDataPoolsOperations - :ivar workspace_git_repo_management: WorkspaceGitRepoManagementOperations operations - :vartype workspace_git_repo_management: azure.synapse.artifacts.v2021_06_01_preview.operations.WorkspaceGitRepoManagementOperations - :ivar integration_runtimes: IntegrationRuntimesOperations operations - :vartype integration_runtimes: azure.synapse.artifacts.v2021_06_01_preview.operations.IntegrationRuntimesOperations - :ivar library: LibraryOperations operations - :vartype library: azure.synapse.artifacts.v2021_06_01_preview.operations.LibraryOperations - :ivar operation_result: OperationResultOperations operations - :vartype operation_result: azure.synapse.artifacts.v2021_06_01_preview.operations.OperationResultOperations - :ivar operation_status: OperationStatusOperations operations - :vartype operation_status: azure.synapse.artifacts.v2021_06_01_preview.operations.OperationStatusOperations - :ivar sql_pools: SqlPoolsOperations operations - :vartype sql_pools: azure.synapse.artifacts.v2021_06_01_preview.operations.SqlPoolsOperations - :ivar workspace: WorkspaceOperations operations - :vartype workspace: azure.synapse.artifacts.v2021_06_01_preview.operations.WorkspaceOperations - :param credential: Credential needed for the client to connect to Azure. - :type credential: ~azure.core.credentials.TokenCredential - :param endpoint: The workspace development endpoint, for example https://myworkspace.dev.azuresynapse.net. - :type endpoint: str - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - """ - - def __init__( - self, - credential, # type: "TokenCredential" - endpoint, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - base_url = '{endpoint}' - self._config = ArtifactsClientConfiguration(credential, endpoint, **kwargs) - self._client = PipelineClient(base_url=base_url, config=self._config, **kwargs) - - client_models = {k: v for k, v in models.__dict__.items() if isinstance(v, type)} - self._serialize = Serializer(client_models) - self._serialize.client_side_validation = False - self._deserialize = Deserializer(client_models) - - self.linked_service = LinkedServiceOperations( - self._client, self._config, self._serialize, self._deserialize) - self.dataset = DatasetOperations( - self._client, self._config, self._serialize, self._deserialize) - self.pipeline = PipelineOperations( - self._client, self._config, self._serialize, self._deserialize) - self.pipeline_run = PipelineRunOperations( - self._client, self._config, self._serialize, self._deserialize) - self.trigger = TriggerOperations( - self._client, self._config, self._serialize, self._deserialize) - self.trigger_run = TriggerRunOperations( - self._client, self._config, self._serialize, self._deserialize) - self.data_flow = DataFlowOperations( - self._client, self._config, self._serialize, self._deserialize) - self.data_flow_debug_session = DataFlowDebugSessionOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_script = SqlScriptOperations( - self._client, self._config, self._serialize, self._deserialize) - self.spark_job_definition = SparkJobDefinitionOperations( - self._client, self._config, self._serialize, self._deserialize) - self.notebook = NotebookOperations( - self._client, self._config, self._serialize, self._deserialize) - self.notebook_operation_result = NotebookOperationResultOperations( - self._client, self._config, self._serialize, self._deserialize) - self.spark_configuration = SparkConfigurationOperations( - self._client, self._config, self._serialize, self._deserialize) - self.big_data_pools = BigDataPoolsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.workspace_git_repo_management = WorkspaceGitRepoManagementOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtimes = IntegrationRuntimesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.library = LibraryOperations( - self._client, self._config, self._serialize, self._deserialize) - self.operation_result = OperationResultOperations( - self._client, self._config, self._serialize, self._deserialize) - self.operation_status = OperationStatusOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pools = SqlPoolsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.workspace = WorkspaceOperations( - self._client, self._config, self._serialize, self._deserialize) - - def _send_request(self, http_request, **kwargs): - # type: (HttpRequest, Any) -> HttpResponse - """Runs the network request through the client's chained policies. - - :param http_request: The network request you want to make. Required. - :type http_request: ~azure.core.pipeline.transport.HttpRequest - :keyword bool stream: Whether the response payload will be streamed. Defaults to True. - :return: The response of your network call. Does not do error handling on your response. - :rtype: ~azure.core.pipeline.transport.HttpResponse - """ - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - http_request.url = self._client.format_url(http_request.url, **path_format_arguments) - stream = kwargs.pop("stream", True) - pipeline_response = self._client._pipeline.run(http_request, stream=stream, **kwargs) - return pipeline_response.http_response - - def close(self): - # type: () -> None - self._client.close() - - def __enter__(self): - # type: () -> ArtifactsClient - self._client.__enter__() - return self - - def __exit__(self, *exc_details): - # type: (Any) -> None - self._client.__exit__(*exc_details) diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/_configuration.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/_configuration.py deleted file mode 100644 index a35193d03a3f..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/_configuration.py +++ /dev/null @@ -1,70 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from typing import TYPE_CHECKING - -from azure.core.configuration import Configuration -from azure.core.pipeline import policies - -from ._version import VERSION - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any - - from azure.core.credentials import TokenCredential - - -class ArtifactsClientConfiguration(Configuration): - """Configuration for ArtifactsClient. - - Note that all parameters used to create this instance are saved as instance - attributes. - - :param credential: Credential needed for the client to connect to Azure. - :type credential: ~azure.core.credentials.TokenCredential - :param endpoint: The workspace development endpoint, for example https://myworkspace.dev.azuresynapse.net. - :type endpoint: str - """ - - def __init__( - self, - credential, # type: "TokenCredential" - endpoint, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - if credential is None: - raise ValueError("Parameter 'credential' must not be None.") - if endpoint is None: - raise ValueError("Parameter 'endpoint' must not be None.") - super(ArtifactsClientConfiguration, self).__init__(**kwargs) - - self.credential = credential - self.endpoint = endpoint - self.api_version = "2021-06-01-preview" - self.credential_scopes = kwargs.pop('credential_scopes', ['https://dev.azuresynapse.net/.default']) - kwargs.setdefault('sdk_moniker', 'synapse-artifacts/{}'.format(VERSION)) - self._configure(**kwargs) - - def _configure( - self, - **kwargs # type: Any - ): - # type: (...) -> None - self.user_agent_policy = kwargs.get('user_agent_policy') or policies.UserAgentPolicy(**kwargs) - self.headers_policy = kwargs.get('headers_policy') or policies.HeadersPolicy(**kwargs) - self.proxy_policy = kwargs.get('proxy_policy') or policies.ProxyPolicy(**kwargs) - self.logging_policy = kwargs.get('logging_policy') or policies.NetworkTraceLoggingPolicy(**kwargs) - self.http_logging_policy = kwargs.get('http_logging_policy') or policies.HttpLoggingPolicy(**kwargs) - self.retry_policy = kwargs.get('retry_policy') or policies.RetryPolicy(**kwargs) - self.custom_hook_policy = kwargs.get('custom_hook_policy') or policies.CustomHookPolicy(**kwargs) - self.redirect_policy = kwargs.get('redirect_policy') or policies.RedirectPolicy(**kwargs) - self.authentication_policy = kwargs.get('authentication_policy') - if self.credential and not self.authentication_policy: - self.authentication_policy = policies.BearerTokenCredentialPolicy(self.credential, *self.credential_scopes, **kwargs) diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/_metadata.json b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/_metadata.json deleted file mode 100644 index ff9ae339af0c..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/_metadata.json +++ /dev/null @@ -1,111 +0,0 @@ -{ - "chosen_version": "2021-06-01-preview", - "total_api_version_list": ["2021-06-01-preview"], - "client": { - "name": "ArtifactsClient", - "filename": "_artifacts_client", - "description": "ArtifactsClient.", - "base_url": null, - "custom_base_url": "\u0027{endpoint}\u0027", - "azure_arm": false, - "has_lro_operations": true, - "client_side_validation": false, - "sync_imports": "{\"typing\": {\"azurecore\": {\"azure.core.credentials\": [\"TokenCredential\"]}}, \"regular\": {\"azurecore\": {\"azure.profiles\": [\"KnownProfiles\", \"ProfileDefinition\"], \"azure.profiles.multiapiclient\": [\"MultiApiClientMixin\"], \"msrest\": [\"Deserializer\", \"Serializer\"], \"azure.core\": [\"PipelineClient\"]}, \"local\": {\"._configuration\": [\"ArtifactsClientConfiguration\"]}}, \"conditional\": {\"stdlib\": {\"typing\": [\"Any\", \"Optional\"]}, \"azurecore\": {\"azure.core.pipeline.transport\": [\"HttpRequest\", \"HttpResponse\"]}}}", - "async_imports": "{\"typing\": {\"azurecore\": {\"azure.core.credentials_async\": [\"AsyncTokenCredential\"]}}, \"regular\": {\"azurecore\": {\"azure.profiles\": [\"KnownProfiles\", \"ProfileDefinition\"], \"azure.profiles.multiapiclient\": [\"MultiApiClientMixin\"], \"msrest\": [\"Deserializer\", \"Serializer\"], \"azure.core\": [\"AsyncPipelineClient\"]}, \"local\": {\"._configuration\": [\"ArtifactsClientConfiguration\"]}}, \"conditional\": {\"stdlib\": {\"typing\": [\"Any\", \"Optional\"]}, \"azurecore\": {\"azure.core.pipeline.transport\": [\"AsyncHttpResponse\", \"HttpRequest\"]}}}" - }, - "global_parameters": { - "sync": { - "credential": { - "signature": "credential, # type: \"TokenCredential\"", - "description": "Credential needed for the client to connect to Azure.", - "docstring_type": "~azure.core.credentials.TokenCredential", - "required": true - }, - "endpoint": { - "signature": "endpoint, # type: str", - "description": "The workspace development endpoint, for example https://myworkspace.dev.azuresynapse.net.", - "docstring_type": "str", - "required": true - } - }, - "async": { - "credential": { - "signature": "credential: \"AsyncTokenCredential\",", - "description": "Credential needed for the client to connect to Azure.", - "docstring_type": "~azure.core.credentials_async.AsyncTokenCredential", - "required": true - }, - "endpoint": { - "signature": "endpoint: str,", - "description": "The workspace development endpoint, for example https://myworkspace.dev.azuresynapse.net.", - "docstring_type": "str", - "required": true - } - }, - "constant": { - }, - "call": "credential, endpoint", - "service_client_specific": { - "sync": { - "api_version": { - "signature": "api_version=None, # type: Optional[str]", - "description": "API version to use if no profile is provided, or if missing in profile.", - "docstring_type": "str", - "required": false - }, - "profile": { - "signature": "profile=KnownProfiles.default, # type: KnownProfiles", - "description": "A profile definition, from KnownProfiles to dict.", - "docstring_type": "azure.profiles.KnownProfiles", - "required": false - } - }, - "async": { - "api_version": { - "signature": "api_version: Optional[str] = None,", - "description": "API version to use if no profile is provided, or if missing in profile.", - "docstring_type": "str", - "required": false - }, - "profile": { - "signature": "profile: KnownProfiles = KnownProfiles.default,", - "description": "A profile definition, from KnownProfiles to dict.", - "docstring_type": "azure.profiles.KnownProfiles", - "required": false - } - } - } - }, - "config": { - "credential": true, - "credential_scopes": ["https://dev.azuresynapse.net/.default"], - "credential_default_policy_type": "BearerTokenCredentialPolicy", - "credential_default_policy_type_has_async_version": true, - "credential_key_header_name": null, - "sync_imports": "{\"regular\": {\"azurecore\": {\"azure.core.configuration\": [\"Configuration\"], \"azure.core.pipeline\": [\"policies\"]}, \"local\": {\"._version\": [\"VERSION\"]}}, \"conditional\": {\"stdlib\": {\"typing\": [\"Any\"]}}, \"typing\": {\"azurecore\": {\"azure.core.credentials\": [\"TokenCredential\"]}}}", - "async_imports": "{\"regular\": {\"azurecore\": {\"azure.core.configuration\": [\"Configuration\"], \"azure.core.pipeline\": [\"policies\"]}, \"local\": {\".._version\": [\"VERSION\"]}}, \"conditional\": {\"stdlib\": {\"typing\": [\"Any\"]}}, \"typing\": {\"azurecore\": {\"azure.core.credentials_async\": [\"AsyncTokenCredential\"]}}}" - }, - "operation_groups": { - "linked_service": "LinkedServiceOperations", - "dataset": "DatasetOperations", - "pipeline": "PipelineOperations", - "pipeline_run": "PipelineRunOperations", - "trigger": "TriggerOperations", - "trigger_run": "TriggerRunOperations", - "data_flow": "DataFlowOperations", - "data_flow_debug_session": "DataFlowDebugSessionOperations", - "sql_script": "SqlScriptOperations", - "spark_job_definition": "SparkJobDefinitionOperations", - "notebook": "NotebookOperations", - "notebook_operation_result": "NotebookOperationResultOperations", - "spark_configuration": "SparkConfigurationOperations", - "big_data_pools": "BigDataPoolsOperations", - "workspace_git_repo_management": "WorkspaceGitRepoManagementOperations", - "integration_runtimes": "IntegrationRuntimesOperations", - "library": "LibraryOperations", - "operation_result": "OperationResultOperations", - "operation_status": "OperationStatusOperations", - "sql_pools": "SqlPoolsOperations", - "workspace": "WorkspaceOperations" - } -} \ No newline at end of file diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/_version.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/_version.py deleted file mode 100644 index eae7c95b6fbd..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/_version.py +++ /dev/null @@ -1,9 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -VERSION = "0.1.0" diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/__init__.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/__init__.py deleted file mode 100644 index 16f882fa0a48..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/__init__.py +++ /dev/null @@ -1,10 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from ._artifacts_client import ArtifactsClient -__all__ = ['ArtifactsClient'] diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/_artifacts_client.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/_artifacts_client.py deleted file mode 100644 index b09ea28b9c85..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/_artifacts_client.py +++ /dev/null @@ -1,179 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from typing import Any, TYPE_CHECKING - -from azure.core import AsyncPipelineClient -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest -from msrest import Deserializer, Serializer - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from azure.core.credentials_async import AsyncTokenCredential - -from ._configuration import ArtifactsClientConfiguration -from .operations import LinkedServiceOperations -from .operations import DatasetOperations -from .operations import PipelineOperations -from .operations import PipelineRunOperations -from .operations import TriggerOperations -from .operations import TriggerRunOperations -from .operations import DataFlowOperations -from .operations import DataFlowDebugSessionOperations -from .operations import SqlScriptOperations -from .operations import SparkJobDefinitionOperations -from .operations import NotebookOperations -from .operations import NotebookOperationResultOperations -from .operations import SparkConfigurationOperations -from .operations import BigDataPoolsOperations -from .operations import WorkspaceGitRepoManagementOperations -from .operations import IntegrationRuntimesOperations -from .operations import LibraryOperations -from .operations import OperationResultOperations -from .operations import OperationStatusOperations -from .operations import SqlPoolsOperations -from .operations import WorkspaceOperations -from .. import models - - -class ArtifactsClient(object): - """ArtifactsClient. - - :ivar linked_service: LinkedServiceOperations operations - :vartype linked_service: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.LinkedServiceOperations - :ivar dataset: DatasetOperations operations - :vartype dataset: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.DatasetOperations - :ivar pipeline: PipelineOperations operations - :vartype pipeline: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.PipelineOperations - :ivar pipeline_run: PipelineRunOperations operations - :vartype pipeline_run: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.PipelineRunOperations - :ivar trigger: TriggerOperations operations - :vartype trigger: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.TriggerOperations - :ivar trigger_run: TriggerRunOperations operations - :vartype trigger_run: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.TriggerRunOperations - :ivar data_flow: DataFlowOperations operations - :vartype data_flow: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.DataFlowOperations - :ivar data_flow_debug_session: DataFlowDebugSessionOperations operations - :vartype data_flow_debug_session: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.DataFlowDebugSessionOperations - :ivar sql_script: SqlScriptOperations operations - :vartype sql_script: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.SqlScriptOperations - :ivar spark_job_definition: SparkJobDefinitionOperations operations - :vartype spark_job_definition: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.SparkJobDefinitionOperations - :ivar notebook: NotebookOperations operations - :vartype notebook: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.NotebookOperations - :ivar notebook_operation_result: NotebookOperationResultOperations operations - :vartype notebook_operation_result: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.NotebookOperationResultOperations - :ivar spark_configuration: SparkConfigurationOperations operations - :vartype spark_configuration: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.SparkConfigurationOperations - :ivar big_data_pools: BigDataPoolsOperations operations - :vartype big_data_pools: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.BigDataPoolsOperations - :ivar workspace_git_repo_management: WorkspaceGitRepoManagementOperations operations - :vartype workspace_git_repo_management: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.WorkspaceGitRepoManagementOperations - :ivar integration_runtimes: IntegrationRuntimesOperations operations - :vartype integration_runtimes: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.IntegrationRuntimesOperations - :ivar library: LibraryOperations operations - :vartype library: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.LibraryOperations - :ivar operation_result: OperationResultOperations operations - :vartype operation_result: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.OperationResultOperations - :ivar operation_status: OperationStatusOperations operations - :vartype operation_status: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.OperationStatusOperations - :ivar sql_pools: SqlPoolsOperations operations - :vartype sql_pools: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.SqlPoolsOperations - :ivar workspace: WorkspaceOperations operations - :vartype workspace: azure.synapse.artifacts.v2021_06_01_preview.aio.operations.WorkspaceOperations - :param credential: Credential needed for the client to connect to Azure. - :type credential: ~azure.core.credentials_async.AsyncTokenCredential - :param endpoint: The workspace development endpoint, for example https://myworkspace.dev.azuresynapse.net. - :type endpoint: str - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - """ - - def __init__( - self, - credential: "AsyncTokenCredential", - endpoint: str, - **kwargs: Any - ) -> None: - base_url = '{endpoint}' - self._config = ArtifactsClientConfiguration(credential, endpoint, **kwargs) - self._client = AsyncPipelineClient(base_url=base_url, config=self._config, **kwargs) - - client_models = {k: v for k, v in models.__dict__.items() if isinstance(v, type)} - self._serialize = Serializer(client_models) - self._serialize.client_side_validation = False - self._deserialize = Deserializer(client_models) - - self.linked_service = LinkedServiceOperations( - self._client, self._config, self._serialize, self._deserialize) - self.dataset = DatasetOperations( - self._client, self._config, self._serialize, self._deserialize) - self.pipeline = PipelineOperations( - self._client, self._config, self._serialize, self._deserialize) - self.pipeline_run = PipelineRunOperations( - self._client, self._config, self._serialize, self._deserialize) - self.trigger = TriggerOperations( - self._client, self._config, self._serialize, self._deserialize) - self.trigger_run = TriggerRunOperations( - self._client, self._config, self._serialize, self._deserialize) - self.data_flow = DataFlowOperations( - self._client, self._config, self._serialize, self._deserialize) - self.data_flow_debug_session = DataFlowDebugSessionOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_script = SqlScriptOperations( - self._client, self._config, self._serialize, self._deserialize) - self.spark_job_definition = SparkJobDefinitionOperations( - self._client, self._config, self._serialize, self._deserialize) - self.notebook = NotebookOperations( - self._client, self._config, self._serialize, self._deserialize) - self.notebook_operation_result = NotebookOperationResultOperations( - self._client, self._config, self._serialize, self._deserialize) - self.spark_configuration = SparkConfigurationOperations( - self._client, self._config, self._serialize, self._deserialize) - self.big_data_pools = BigDataPoolsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.workspace_git_repo_management = WorkspaceGitRepoManagementOperations( - self._client, self._config, self._serialize, self._deserialize) - self.integration_runtimes = IntegrationRuntimesOperations( - self._client, self._config, self._serialize, self._deserialize) - self.library = LibraryOperations( - self._client, self._config, self._serialize, self._deserialize) - self.operation_result = OperationResultOperations( - self._client, self._config, self._serialize, self._deserialize) - self.operation_status = OperationStatusOperations( - self._client, self._config, self._serialize, self._deserialize) - self.sql_pools = SqlPoolsOperations( - self._client, self._config, self._serialize, self._deserialize) - self.workspace = WorkspaceOperations( - self._client, self._config, self._serialize, self._deserialize) - - async def _send_request(self, http_request: HttpRequest, **kwargs: Any) -> AsyncHttpResponse: - """Runs the network request through the client's chained policies. - - :param http_request: The network request you want to make. Required. - :type http_request: ~azure.core.pipeline.transport.HttpRequest - :keyword bool stream: Whether the response payload will be streamed. Defaults to True. - :return: The response of your network call. Does not do error handling on your response. - :rtype: ~azure.core.pipeline.transport.AsyncHttpResponse - """ - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - http_request.url = self._client.format_url(http_request.url, **path_format_arguments) - stream = kwargs.pop("stream", True) - pipeline_response = await self._client._pipeline.run(http_request, stream=stream, **kwargs) - return pipeline_response.http_response - - async def close(self) -> None: - await self._client.close() - - async def __aenter__(self) -> "ArtifactsClient": - await self._client.__aenter__() - return self - - async def __aexit__(self, *exc_details) -> None: - await self._client.__aexit__(*exc_details) diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/_configuration.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/_configuration.py deleted file mode 100644 index f7ab4ca7b235..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/_configuration.py +++ /dev/null @@ -1,66 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from typing import Any, TYPE_CHECKING - -from azure.core.configuration import Configuration -from azure.core.pipeline import policies - -from .._version import VERSION - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from azure.core.credentials_async import AsyncTokenCredential - - -class ArtifactsClientConfiguration(Configuration): - """Configuration for ArtifactsClient. - - Note that all parameters used to create this instance are saved as instance - attributes. - - :param credential: Credential needed for the client to connect to Azure. - :type credential: ~azure.core.credentials_async.AsyncTokenCredential - :param endpoint: The workspace development endpoint, for example https://myworkspace.dev.azuresynapse.net. - :type endpoint: str - """ - - def __init__( - self, - credential: "AsyncTokenCredential", - endpoint: str, - **kwargs: Any - ) -> None: - if credential is None: - raise ValueError("Parameter 'credential' must not be None.") - if endpoint is None: - raise ValueError("Parameter 'endpoint' must not be None.") - super(ArtifactsClientConfiguration, self).__init__(**kwargs) - - self.credential = credential - self.endpoint = endpoint - self.api_version = "2021-06-01-preview" - self.credential_scopes = kwargs.pop('credential_scopes', ['https://dev.azuresynapse.net/.default']) - kwargs.setdefault('sdk_moniker', 'synapse-artifacts/{}'.format(VERSION)) - self._configure(**kwargs) - - def _configure( - self, - **kwargs: Any - ) -> None: - self.user_agent_policy = kwargs.get('user_agent_policy') or policies.UserAgentPolicy(**kwargs) - self.headers_policy = kwargs.get('headers_policy') or policies.HeadersPolicy(**kwargs) - self.proxy_policy = kwargs.get('proxy_policy') or policies.ProxyPolicy(**kwargs) - self.logging_policy = kwargs.get('logging_policy') or policies.NetworkTraceLoggingPolicy(**kwargs) - self.http_logging_policy = kwargs.get('http_logging_policy') or policies.HttpLoggingPolicy(**kwargs) - self.retry_policy = kwargs.get('retry_policy') or policies.AsyncRetryPolicy(**kwargs) - self.custom_hook_policy = kwargs.get('custom_hook_policy') or policies.CustomHookPolicy(**kwargs) - self.redirect_policy = kwargs.get('redirect_policy') or policies.AsyncRedirectPolicy(**kwargs) - self.authentication_policy = kwargs.get('authentication_policy') - if self.credential and not self.authentication_policy: - self.authentication_policy = policies.AsyncBearerTokenCredentialPolicy(self.credential, *self.credential_scopes, **kwargs) diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_big_data_pools_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_big_data_pools_operations.py deleted file mode 100644 index 96eb296c471e..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_big_data_pools_operations.py +++ /dev/null @@ -1,146 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, Callable, Dict, Generic, Optional, TypeVar -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class BigDataPoolsOperations: - """BigDataPoolsOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - async def list( - self, - **kwargs: Any - ) -> "_models.BigDataPoolResourceInfoListResult": - """List Big Data Pools. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: BigDataPoolResourceInfoListResult, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.BigDataPoolResourceInfoListResult - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.BigDataPoolResourceInfoListResult"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.list.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('BigDataPoolResourceInfoListResult', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - list.metadata = {'url': '/bigDataPools'} # type: ignore - - async def get( - self, - big_data_pool_name: str, - **kwargs: Any - ) -> "_models.BigDataPoolResourceInfo": - """Get Big Data Pool. - - :param big_data_pool_name: The Big Data Pool name. - :type big_data_pool_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: BigDataPoolResourceInfo, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.BigDataPoolResourceInfo - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.BigDataPoolResourceInfo"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'bigDataPoolName': self._serialize.url("big_data_pool_name", big_data_pool_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('BigDataPoolResourceInfo', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get.metadata = {'url': '/bigDataPools/{bigDataPoolName}'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_data_flow_debug_session_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_data_flow_debug_session_operations.py deleted file mode 100644 index 73a84fa24281..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_data_flow_debug_session_operations.py +++ /dev/null @@ -1,463 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, AsyncIterable, Callable, Dict, Generic, Optional, TypeVar, Union -import warnings - -from azure.core.async_paging import AsyncItemPaged, AsyncList -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest -from azure.core.polling import AsyncLROPoller, AsyncNoPolling, AsyncPollingMethod -from azure.core.polling.async_base_polling import AsyncLROBasePolling - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class DataFlowDebugSessionOperations: - """DataFlowDebugSessionOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - async def _create_data_flow_debug_session_initial( - self, - request: "_models.CreateDataFlowDebugSessionRequest", - **kwargs: Any - ) -> Optional["_models.CreateDataFlowDebugSessionResponse"]: - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.CreateDataFlowDebugSessionResponse"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._create_data_flow_debug_session_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(request, 'CreateDataFlowDebugSessionRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - response_headers = {} - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('CreateDataFlowDebugSessionResponse', pipeline_response) - - if response.status_code == 202: - response_headers['location']=self._deserialize('str', response.headers.get('location')) - - if cls: - return cls(pipeline_response, deserialized, response_headers) - - return deserialized - _create_data_flow_debug_session_initial.metadata = {'url': '/createDataFlowDebugSession'} # type: ignore - - async def begin_create_data_flow_debug_session( - self, - request: "_models.CreateDataFlowDebugSessionRequest", - **kwargs: Any - ) -> AsyncLROPoller["_models.CreateDataFlowDebugSessionResponse"]: - """Creates a data flow debug session. - - :param request: Data flow debug session definition. - :type request: ~azure.synapse.artifacts.v2021_06_01_preview.models.CreateDataFlowDebugSessionRequest - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either CreateDataFlowDebugSessionResponse or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.CreateDataFlowDebugSessionResponse] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.CreateDataFlowDebugSessionResponse"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._create_data_flow_debug_session_initial( - request=request, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('CreateDataFlowDebugSessionResponse', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create_data_flow_debug_session.metadata = {'url': '/createDataFlowDebugSession'} # type: ignore - - def query_data_flow_debug_sessions_by_workspace( - self, - **kwargs: Any - ) -> AsyncIterable["_models.QueryDataFlowDebugSessionsResponse"]: - """Query all active data flow debug sessions. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either QueryDataFlowDebugSessionsResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.QueryDataFlowDebugSessionsResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.QueryDataFlowDebugSessionsResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.query_data_flow_debug_sessions_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - async def extract_data(pipeline_response): - deserialized = self._deserialize('QueryDataFlowDebugSessionsResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, AsyncList(list_of_elem) - - async def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return AsyncItemPaged( - get_next, extract_data - ) - query_data_flow_debug_sessions_by_workspace.metadata = {'url': '/queryDataFlowDebugSessions'} # type: ignore - - async def add_data_flow( - self, - request: "_models.DataFlowDebugPackage", - **kwargs: Any - ) -> "_models.AddDataFlowToDebugSessionResponse": - """Add a data flow into debug session. - - :param request: Data flow debug session definition with debug content. - :type request: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowDebugPackage - :keyword callable cls: A custom type or function that will be passed the direct response - :return: AddDataFlowToDebugSessionResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.AddDataFlowToDebugSessionResponse - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.AddDataFlowToDebugSessionResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self.add_data_flow.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(request, 'DataFlowDebugPackage') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('AddDataFlowToDebugSessionResponse', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - add_data_flow.metadata = {'url': '/addDataFlowToDebugSession'} # type: ignore - - async def delete_data_flow_debug_session( - self, - request: "_models.DeleteDataFlowDebugSessionRequest", - **kwargs: Any - ) -> None: - """Deletes a data flow debug session. - - :param request: Data flow debug session definition for deletion. - :type request: ~azure.synapse.artifacts.v2021_06_01_preview.models.DeleteDataFlowDebugSessionRequest - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self.delete_data_flow_debug_session.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(request, 'DeleteDataFlowDebugSessionRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - delete_data_flow_debug_session.metadata = {'url': '/deleteDataFlowDebugSession'} # type: ignore - - async def _execute_command_initial( - self, - request: "_models.DataFlowDebugCommandRequest", - **kwargs: Any - ) -> Optional["_models.DataFlowDebugCommandResponse"]: - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.DataFlowDebugCommandResponse"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._execute_command_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(request, 'DataFlowDebugCommandRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - response_headers = {} - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('DataFlowDebugCommandResponse', pipeline_response) - - if response.status_code == 202: - response_headers['location']=self._deserialize('str', response.headers.get('location')) - - if cls: - return cls(pipeline_response, deserialized, response_headers) - - return deserialized - _execute_command_initial.metadata = {'url': '/executeDataFlowDebugCommand'} # type: ignore - - async def begin_execute_command( - self, - request: "_models.DataFlowDebugCommandRequest", - **kwargs: Any - ) -> AsyncLROPoller["_models.DataFlowDebugCommandResponse"]: - """Execute a data flow debug command. - - :param request: Data flow debug command definition. - :type request: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowDebugCommandRequest - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either DataFlowDebugCommandResponse or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowDebugCommandResponse] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.DataFlowDebugCommandResponse"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._execute_command_initial( - request=request, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('DataFlowDebugCommandResponse', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_execute_command.metadata = {'url': '/executeDataFlowDebugCommand'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_data_flow_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_data_flow_operations.py deleted file mode 100644 index e3d083b76d02..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_data_flow_operations.py +++ /dev/null @@ -1,521 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, AsyncIterable, Callable, Dict, Generic, Optional, TypeVar, Union -import warnings - -from azure.core.async_paging import AsyncItemPaged, AsyncList -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest -from azure.core.polling import AsyncLROPoller, AsyncNoPolling, AsyncPollingMethod -from azure.core.polling.async_base_polling import AsyncLROBasePolling - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class DataFlowOperations: - """DataFlowOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - async def _create_or_update_data_flow_initial( - self, - data_flow_name: str, - properties: "_models.DataFlow", - if_match: Optional[str] = None, - **kwargs: Any - ) -> Optional["_models.DataFlowResource"]: - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.DataFlowResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _data_flow = _models.DataFlowResource(properties=properties) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._create_or_update_data_flow_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'dataFlowName': self._serialize.url("data_flow_name", data_flow_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_match is not None: - header_parameters['If-Match'] = self._serialize.header("if_match", if_match, 'str') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_data_flow, 'DataFlowResource') - body_content_kwargs['content'] = body_content - request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('DataFlowResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _create_or_update_data_flow_initial.metadata = {'url': '/dataflows/{dataFlowName}'} # type: ignore - - async def begin_create_or_update_data_flow( - self, - data_flow_name: str, - properties: "_models.DataFlow", - if_match: Optional[str] = None, - **kwargs: Any - ) -> AsyncLROPoller["_models.DataFlowResource"]: - """Creates or updates a data flow. - - :param data_flow_name: The data flow name. - :type data_flow_name: str - :param properties: Data flow properties. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlow - :param if_match: ETag of the data flow entity. Should only be specified for update, for which - it should match existing entity or can be * for unconditional update. - :type if_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either DataFlowResource or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowResource] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.DataFlowResource"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._create_or_update_data_flow_initial( - data_flow_name=data_flow_name, - properties=properties, - if_match=if_match, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('DataFlowResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'dataFlowName': self._serialize.url("data_flow_name", data_flow_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create_or_update_data_flow.metadata = {'url': '/dataflows/{dataFlowName}'} # type: ignore - - async def get_data_flow( - self, - data_flow_name: str, - if_none_match: Optional[str] = None, - **kwargs: Any - ) -> "_models.DataFlowResource": - """Gets a data flow. - - :param data_flow_name: The data flow name. - :type data_flow_name: str - :param if_none_match: ETag of the data flow entity. Should only be specified for get. If the - ETag matches the existing entity tag, or if * was provided, then no content will be returned. - :type if_none_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: DataFlowResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowResource - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.DataFlowResource"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_data_flow.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'dataFlowName': self._serialize.url("data_flow_name", data_flow_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_none_match is not None: - header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('DataFlowResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_data_flow.metadata = {'url': '/dataflows/{dataFlowName}'} # type: ignore - - async def _delete_data_flow_initial( - self, - data_flow_name: str, - **kwargs: Any - ) -> None: - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._delete_data_flow_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'dataFlowName': self._serialize.url("data_flow_name", data_flow_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.delete(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _delete_data_flow_initial.metadata = {'url': '/dataflows/{dataFlowName}'} # type: ignore - - async def begin_delete_data_flow( - self, - data_flow_name: str, - **kwargs: Any - ) -> AsyncLROPoller[None]: - """Deletes a data flow. - - :param data_flow_name: The data flow name. - :type data_flow_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._delete_data_flow_initial( - data_flow_name=data_flow_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'dataFlowName': self._serialize.url("data_flow_name", data_flow_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_delete_data_flow.metadata = {'url': '/dataflows/{dataFlowName}'} # type: ignore - - async def _rename_data_flow_initial( - self, - data_flow_name: str, - new_name: Optional[str] = None, - **kwargs: Any - ) -> None: - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _request = _models.ArtifactRenameRequest(new_name=new_name) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._rename_data_flow_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'dataFlowName': self._serialize.url("data_flow_name", data_flow_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_request, 'ArtifactRenameRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _rename_data_flow_initial.metadata = {'url': '/dataflows/{dataFlowName}/rename'} # type: ignore - - async def begin_rename_data_flow( - self, - data_flow_name: str, - new_name: Optional[str] = None, - **kwargs: Any - ) -> AsyncLROPoller[None]: - """Renames a dataflow. - - :param data_flow_name: The data flow name. - :type data_flow_name: str - :param new_name: New name of the artifact. - :type new_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._rename_data_flow_initial( - data_flow_name=data_flow_name, - new_name=new_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'dataFlowName': self._serialize.url("data_flow_name", data_flow_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_rename_data_flow.metadata = {'url': '/dataflows/{dataFlowName}/rename'} # type: ignore - - def get_data_flows_by_workspace( - self, - **kwargs: Any - ) -> AsyncIterable["_models.DataFlowListResponse"]: - """Lists data flows. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either DataFlowListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.DataFlowListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.get_data_flows_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - async def extract_data(pipeline_response): - deserialized = self._deserialize('DataFlowListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, AsyncList(list_of_elem) - - async def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return AsyncItemPaged( - get_next, extract_data - ) - get_data_flows_by_workspace.metadata = {'url': '/dataflows'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_dataset_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_dataset_operations.py deleted file mode 100644 index bf32a709c7a6..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_dataset_operations.py +++ /dev/null @@ -1,523 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, AsyncIterable, Callable, Dict, Generic, Optional, TypeVar, Union -import warnings - -from azure.core.async_paging import AsyncItemPaged, AsyncList -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest -from azure.core.polling import AsyncLROPoller, AsyncNoPolling, AsyncPollingMethod -from azure.core.polling.async_base_polling import AsyncLROBasePolling - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class DatasetOperations: - """DatasetOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get_datasets_by_workspace( - self, - **kwargs: Any - ) -> AsyncIterable["_models.DatasetListResponse"]: - """Lists datasets. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either DatasetListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.DatasetListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.get_datasets_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - async def extract_data(pipeline_response): - deserialized = self._deserialize('DatasetListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, AsyncList(list_of_elem) - - async def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return AsyncItemPaged( - get_next, extract_data - ) - get_datasets_by_workspace.metadata = {'url': '/datasets'} # type: ignore - - async def _create_or_update_dataset_initial( - self, - dataset_name: str, - properties: "_models.Dataset", - if_match: Optional[str] = None, - **kwargs: Any - ) -> Optional["_models.DatasetResource"]: - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.DatasetResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _dataset = _models.DatasetResource(properties=properties) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._create_or_update_dataset_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'datasetName': self._serialize.url("dataset_name", dataset_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_match is not None: - header_parameters['If-Match'] = self._serialize.header("if_match", if_match, 'str') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_dataset, 'DatasetResource') - body_content_kwargs['content'] = body_content - request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('DatasetResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _create_or_update_dataset_initial.metadata = {'url': '/datasets/{datasetName}'} # type: ignore - - async def begin_create_or_update_dataset( - self, - dataset_name: str, - properties: "_models.Dataset", - if_match: Optional[str] = None, - **kwargs: Any - ) -> AsyncLROPoller["_models.DatasetResource"]: - """Creates or updates a dataset. - - :param dataset_name: The dataset name. - :type dataset_name: str - :param properties: Dataset properties. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.Dataset - :param if_match: ETag of the dataset entity. Should only be specified for update, for which it - should match existing entity or can be * for unconditional update. - :type if_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either DatasetResource or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetResource] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.DatasetResource"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._create_or_update_dataset_initial( - dataset_name=dataset_name, - properties=properties, - if_match=if_match, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('DatasetResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'datasetName': self._serialize.url("dataset_name", dataset_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create_or_update_dataset.metadata = {'url': '/datasets/{datasetName}'} # type: ignore - - async def get_dataset( - self, - dataset_name: str, - if_none_match: Optional[str] = None, - **kwargs: Any - ) -> Optional["_models.DatasetResource"]: - """Gets a dataset. - - :param dataset_name: The dataset name. - :type dataset_name: str - :param if_none_match: ETag of the dataset entity. Should only be specified for get. If the ETag - matches the existing entity tag, or if * was provided, then no content will be returned. - :type if_none_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: DatasetResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetResource or None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.DatasetResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_dataset.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'datasetName': self._serialize.url("dataset_name", dataset_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_none_match is not None: - header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 304]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('DatasetResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_dataset.metadata = {'url': '/datasets/{datasetName}'} # type: ignore - - async def _delete_dataset_initial( - self, - dataset_name: str, - **kwargs: Any - ) -> None: - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._delete_dataset_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'datasetName': self._serialize.url("dataset_name", dataset_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.delete(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _delete_dataset_initial.metadata = {'url': '/datasets/{datasetName}'} # type: ignore - - async def begin_delete_dataset( - self, - dataset_name: str, - **kwargs: Any - ) -> AsyncLROPoller[None]: - """Deletes a dataset. - - :param dataset_name: The dataset name. - :type dataset_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._delete_dataset_initial( - dataset_name=dataset_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'datasetName': self._serialize.url("dataset_name", dataset_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_delete_dataset.metadata = {'url': '/datasets/{datasetName}'} # type: ignore - - async def _rename_dataset_initial( - self, - dataset_name: str, - new_name: Optional[str] = None, - **kwargs: Any - ) -> None: - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _request = _models.ArtifactRenameRequest(new_name=new_name) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._rename_dataset_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'datasetName': self._serialize.url("dataset_name", dataset_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_request, 'ArtifactRenameRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _rename_dataset_initial.metadata = {'url': '/datasets/{datasetName}/rename'} # type: ignore - - async def begin_rename_dataset( - self, - dataset_name: str, - new_name: Optional[str] = None, - **kwargs: Any - ) -> AsyncLROPoller[None]: - """Renames a dataset. - - :param dataset_name: The dataset name. - :type dataset_name: str - :param new_name: New name of the artifact. - :type new_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._rename_dataset_initial( - dataset_name=dataset_name, - new_name=new_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'datasetName': self._serialize.url("dataset_name", dataset_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_rename_dataset.metadata = {'url': '/datasets/{datasetName}/rename'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_integration_runtimes_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_integration_runtimes_operations.py deleted file mode 100644 index 8567621320fb..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_integration_runtimes_operations.py +++ /dev/null @@ -1,146 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, Callable, Dict, Generic, Optional, TypeVar -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class IntegrationRuntimesOperations: - """IntegrationRuntimesOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - async def list( - self, - **kwargs: Any - ) -> "_models.IntegrationRuntimeListResponse": - """List Integration Runtimes. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: IntegrationRuntimeListResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeListResponse - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.IntegrationRuntimeListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.list.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('IntegrationRuntimeListResponse', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - list.metadata = {'url': '/integrationRuntimes'} # type: ignore - - async def get( - self, - integration_runtime_name: str, - **kwargs: Any - ) -> "_models.IntegrationRuntimeResource": - """Get Integration Runtime. - - :param integration_runtime_name: The Integration Runtime name. - :type integration_runtime_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: IntegrationRuntimeResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeResource - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.IntegrationRuntimeResource"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'integrationRuntimeName': self._serialize.url("integration_runtime_name", integration_runtime_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('IntegrationRuntimeResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get.metadata = {'url': '/integrationRuntimes/{integrationRuntimeName}'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_library_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_library_operations.py deleted file mode 100644 index eacf764ec309..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_library_operations.py +++ /dev/null @@ -1,635 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, AsyncIterable, Callable, Dict, Generic, IO, Optional, TypeVar, Union -import warnings - -from azure.core.async_paging import AsyncItemPaged, AsyncList -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest -from azure.core.polling import AsyncLROPoller, AsyncNoPolling, AsyncPollingMethod -from azure.core.polling.async_base_polling import AsyncLROBasePolling - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class LibraryOperations: - """LibraryOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def list( - self, - **kwargs: Any - ) -> AsyncIterable["_models.LibraryListResponse"]: - """Lists Library. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either LibraryListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.LibraryListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.LibraryListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.list.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - async def extract_data(pipeline_response): - deserialized = self._deserialize('LibraryListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, AsyncList(list_of_elem) - - async def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return AsyncItemPaged( - get_next, extract_data - ) - list.metadata = {'url': '/libraries'} # type: ignore - - async def _flush_initial( - self, - library_name: str, - **kwargs: Any - ) -> Optional["_models.LibraryResourceInfo"]: - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.LibraryResourceInfo"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._flush_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'libraryName': self._serialize.url("library_name", library_name, 'str', max_length=100, min_length=0), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 202: - deserialized = self._deserialize('LibraryResourceInfo', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _flush_initial.metadata = {'url': '/libraries/{libraryName}/flush'} # type: ignore - - async def begin_flush( - self, - library_name: str, - **kwargs: Any - ) -> AsyncLROPoller["_models.LibraryResourceInfo"]: - """Flush Library. - - :param library_name: file name to upload. Minimum length of the filename should be 1 excluding - the extension length. - :type library_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.LibraryResourceInfo"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._flush_initial( - library_name=library_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('LibraryResourceInfo', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'libraryName': self._serialize.url("library_name", library_name, 'str', max_length=100, min_length=0), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_flush.metadata = {'url': '/libraries/{libraryName}/flush'} # type: ignore - - async def get_operation_result( - self, - operation_id: str, - **kwargs: Any - ) -> Union["_models.LibraryResource", "_models.OperationResult"]: - """Get Operation result for Library. - - :param operation_id: operation id for which status is requested. - :type operation_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: LibraryResource or OperationResult, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.LibraryResource or ~azure.synapse.artifacts.v2021_06_01_preview.models.OperationResult - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[Union["_models.LibraryResource", "_models.OperationResult"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_operation_result.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'operationId': self._serialize.url("operation_id", operation_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if response.status_code == 200: - deserialized = self._deserialize('LibraryResource', pipeline_response) - - if response.status_code == 202: - deserialized = self._deserialize('OperationResult', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_operation_result.metadata = {'url': '/libraryOperationResults/{operationId}'} # type: ignore - - async def _delete_initial( - self, - library_name: str, - **kwargs: Any - ) -> Optional["_models.LibraryResourceInfo"]: - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.LibraryResourceInfo"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._delete_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'libraryName': self._serialize.url("library_name", library_name, 'str', max_length=100, min_length=0), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.delete(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202, 409]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 202: - deserialized = self._deserialize('LibraryResourceInfo', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _delete_initial.metadata = {'url': '/libraries/{libraryName}'} # type: ignore - - async def begin_delete( - self, - library_name: str, - **kwargs: Any - ) -> AsyncLROPoller["_models.LibraryResourceInfo"]: - """Delete Library. - - :param library_name: file name to upload. Minimum length of the filename should be 1 excluding - the extension length. - :type library_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.LibraryResourceInfo"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._delete_initial( - library_name=library_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('LibraryResourceInfo', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'libraryName': self._serialize.url("library_name", library_name, 'str', max_length=100, min_length=0), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_delete.metadata = {'url': '/libraries/{libraryName}'} # type: ignore - - async def get( - self, - library_name: str, - **kwargs: Any - ) -> Optional["_models.LibraryResource"]: - """Get Library. - - :param library_name: file name to upload. Minimum length of the filename should be 1 excluding - the extension length. - :type library_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: LibraryResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.LibraryResource or None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.LibraryResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'libraryName': self._serialize.url("library_name", library_name, 'str', max_length=100, min_length=0), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 304]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('LibraryResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get.metadata = {'url': '/libraries/{libraryName}'} # type: ignore - - async def _create_initial( - self, - library_name: str, - **kwargs: Any - ) -> Optional["_models.LibraryResourceInfo"]: - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.LibraryResourceInfo"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._create_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'libraryName': self._serialize.url("library_name", library_name, 'str', max_length=100, min_length=0), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.put(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 202: - deserialized = self._deserialize('LibraryResourceInfo', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _create_initial.metadata = {'url': '/libraries/{libraryName}'} # type: ignore - - async def begin_create( - self, - library_name: str, - **kwargs: Any - ) -> AsyncLROPoller["_models.LibraryResourceInfo"]: - """Creates a library with the library name. - - :param library_name: file name to upload. Minimum length of the filename should be 1 excluding - the extension length. - :type library_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.LibraryResourceInfo"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._create_initial( - library_name=library_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('LibraryResourceInfo', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'libraryName': self._serialize.url("library_name", library_name, 'str', max_length=100, min_length=0), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create.metadata = {'url': '/libraries/{libraryName}'} # type: ignore - - async def append( - self, - library_name: str, - content: IO, - blob_condition_append_position: Optional[int] = None, - **kwargs: Any - ) -> None: - """Append the content to the library resource created using the create operation. The maximum - content size is 4MiB. Content larger than 4MiB must be appended in 4MiB chunks. - - :param library_name: file name to upload. Minimum length of the filename should be 1 excluding - the extension length. - :type library_name: str - :param content: Library file chunk. - :type content: IO - :param blob_condition_append_position: Set this header to a byte offset at which the block is - expected to be appended. The request succeeds only if the current offset matches this value. - Otherwise, the request fails with the AppendPositionConditionNotMet error (HTTP status code 412 - – Precondition Failed). - :type blob_condition_append_position: long - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - comp = "appendblock" - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/octet-stream") - accept = "application/json" - - # Construct URL - url = self.append.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'libraryName': self._serialize.url("library_name", library_name, 'str', max_length=100, min_length=0), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['comp'] = self._serialize.query("comp", comp, 'str') - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if blob_condition_append_position is not None: - header_parameters['x-ms-blob-condition-appendpos'] = self._serialize.header("blob_condition_append_position", blob_condition_append_position, 'long') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content_kwargs['stream_content'] = content - request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [201]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - append.metadata = {'url': '/libraries/{libraryName}'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_linked_service_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_linked_service_operations.py deleted file mode 100644 index 6b0c6f1afbe3..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_linked_service_operations.py +++ /dev/null @@ -1,524 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, AsyncIterable, Callable, Dict, Generic, Optional, TypeVar, Union -import warnings - -from azure.core.async_paging import AsyncItemPaged, AsyncList -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest -from azure.core.polling import AsyncLROPoller, AsyncNoPolling, AsyncPollingMethod -from azure.core.polling.async_base_polling import AsyncLROBasePolling - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class LinkedServiceOperations: - """LinkedServiceOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get_linked_services_by_workspace( - self, - **kwargs: Any - ) -> AsyncIterable["_models.LinkedServiceListResponse"]: - """Lists linked services. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either LinkedServiceListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.LinkedServiceListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.get_linked_services_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - async def extract_data(pipeline_response): - deserialized = self._deserialize('LinkedServiceListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, AsyncList(list_of_elem) - - async def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return AsyncItemPaged( - get_next, extract_data - ) - get_linked_services_by_workspace.metadata = {'url': '/linkedservices'} # type: ignore - - async def _create_or_update_linked_service_initial( - self, - linked_service_name: str, - properties: "_models.LinkedService", - if_match: Optional[str] = None, - **kwargs: Any - ) -> Optional["_models.LinkedServiceResource"]: - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.LinkedServiceResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _linked_service = _models.LinkedServiceResource(properties=properties) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._create_or_update_linked_service_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'linkedServiceName': self._serialize.url("linked_service_name", linked_service_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_match is not None: - header_parameters['If-Match'] = self._serialize.header("if_match", if_match, 'str') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_linked_service, 'LinkedServiceResource') - body_content_kwargs['content'] = body_content - request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('LinkedServiceResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _create_or_update_linked_service_initial.metadata = {'url': '/linkedservices/{linkedServiceName}'} # type: ignore - - async def begin_create_or_update_linked_service( - self, - linked_service_name: str, - properties: "_models.LinkedService", - if_match: Optional[str] = None, - **kwargs: Any - ) -> AsyncLROPoller["_models.LinkedServiceResource"]: - """Creates or updates a linked service. - - :param linked_service_name: The linked service name. - :type linked_service_name: str - :param properties: Properties of linked service. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedService - :param if_match: ETag of the linkedService entity. Should only be specified for update, for - which it should match existing entity or can be * for unconditional update. - :type if_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either LinkedServiceResource or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceResource] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.LinkedServiceResource"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._create_or_update_linked_service_initial( - linked_service_name=linked_service_name, - properties=properties, - if_match=if_match, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('LinkedServiceResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'linkedServiceName': self._serialize.url("linked_service_name", linked_service_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create_or_update_linked_service.metadata = {'url': '/linkedservices/{linkedServiceName}'} # type: ignore - - async def get_linked_service( - self, - linked_service_name: str, - if_none_match: Optional[str] = None, - **kwargs: Any - ) -> Optional["_models.LinkedServiceResource"]: - """Gets a linked service. - - :param linked_service_name: The linked service name. - :type linked_service_name: str - :param if_none_match: ETag of the linked service entity. Should only be specified for get. If - the ETag matches the existing entity tag, or if * was provided, then no content will be - returned. - :type if_none_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: LinkedServiceResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceResource or None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.LinkedServiceResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_linked_service.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'linkedServiceName': self._serialize.url("linked_service_name", linked_service_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_none_match is not None: - header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 304]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('LinkedServiceResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_linked_service.metadata = {'url': '/linkedservices/{linkedServiceName}'} # type: ignore - - async def _delete_linked_service_initial( - self, - linked_service_name: str, - **kwargs: Any - ) -> None: - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._delete_linked_service_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'linkedServiceName': self._serialize.url("linked_service_name", linked_service_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.delete(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _delete_linked_service_initial.metadata = {'url': '/linkedservices/{linkedServiceName}'} # type: ignore - - async def begin_delete_linked_service( - self, - linked_service_name: str, - **kwargs: Any - ) -> AsyncLROPoller[None]: - """Deletes a linked service. - - :param linked_service_name: The linked service name. - :type linked_service_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._delete_linked_service_initial( - linked_service_name=linked_service_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'linkedServiceName': self._serialize.url("linked_service_name", linked_service_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_delete_linked_service.metadata = {'url': '/linkedservices/{linkedServiceName}'} # type: ignore - - async def _rename_linked_service_initial( - self, - linked_service_name: str, - new_name: Optional[str] = None, - **kwargs: Any - ) -> None: - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _request = _models.ArtifactRenameRequest(new_name=new_name) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._rename_linked_service_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'linkedServiceName': self._serialize.url("linked_service_name", linked_service_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_request, 'ArtifactRenameRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _rename_linked_service_initial.metadata = {'url': '/linkedservices/{linkedServiceName}/rename'} # type: ignore - - async def begin_rename_linked_service( - self, - linked_service_name: str, - new_name: Optional[str] = None, - **kwargs: Any - ) -> AsyncLROPoller[None]: - """Renames a linked service. - - :param linked_service_name: The linked service name. - :type linked_service_name: str - :param new_name: New name of the artifact. - :type new_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._rename_linked_service_initial( - linked_service_name=linked_service_name, - new_name=new_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'linkedServiceName': self._serialize.url("linked_service_name", linked_service_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_rename_linked_service.metadata = {'url': '/linkedservices/{linkedServiceName}/rename'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_notebook_operation_result_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_notebook_operation_result_operations.py deleted file mode 100644 index 5146a082c7a0..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_notebook_operation_result_operations.py +++ /dev/null @@ -1,92 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, Callable, Dict, Generic, Optional, TypeVar -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class NotebookOperationResultOperations: - """NotebookOperationResultOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - async def get( - self, - operation_id: str, - **kwargs: Any - ) -> None: - """Get notebook operation result. - - :param operation_id: Operation ID. - :type operation_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'operationId': self._serialize.url("operation_id", operation_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 201, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - get.metadata = {'url': '/notebookOperationResults/{operationId}'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_notebook_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_notebook_operations.py deleted file mode 100644 index 13ac417c45ff..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_notebook_operations.py +++ /dev/null @@ -1,592 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, AsyncIterable, Callable, Dict, Generic, Optional, TypeVar, Union -import warnings - -from azure.core.async_paging import AsyncItemPaged, AsyncList -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest -from azure.core.polling import AsyncLROPoller, AsyncNoPolling, AsyncPollingMethod -from azure.core.polling.async_base_polling import AsyncLROBasePolling - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class NotebookOperations: - """NotebookOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get_notebooks_by_workspace( - self, - **kwargs: Any - ) -> AsyncIterable["_models.NotebookListResponse"]: - """Lists Notebooks. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either NotebookListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.NotebookListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.get_notebooks_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - async def extract_data(pipeline_response): - deserialized = self._deserialize('NotebookListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, AsyncList(list_of_elem) - - async def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return AsyncItemPaged( - get_next, extract_data - ) - get_notebooks_by_workspace.metadata = {'url': '/notebooks'} # type: ignore - - def get_notebook_summary_by_work_space( - self, - **kwargs: Any - ) -> AsyncIterable["_models.NotebookListResponse"]: - """Lists a summary of Notebooks. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either NotebookListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.NotebookListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.get_notebook_summary_by_work_space.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - async def extract_data(pipeline_response): - deserialized = self._deserialize('NotebookListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, AsyncList(list_of_elem) - - async def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return AsyncItemPaged( - get_next, extract_data - ) - get_notebook_summary_by_work_space.metadata = {'url': '/notebooksSummary'} # type: ignore - - async def _create_or_update_notebook_initial( - self, - notebook_name: str, - notebook: "_models.NotebookResource", - if_match: Optional[str] = None, - **kwargs: Any - ) -> Optional["_models.NotebookResource"]: - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.NotebookResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._create_or_update_notebook_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'notebookName': self._serialize.url("notebook_name", notebook_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_match is not None: - header_parameters['If-Match'] = self._serialize.header("if_match", if_match, 'str') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(notebook, 'NotebookResource') - body_content_kwargs['content'] = body_content - request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('NotebookResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _create_or_update_notebook_initial.metadata = {'url': '/notebooks/{notebookName}'} # type: ignore - - async def begin_create_or_update_notebook( - self, - notebook_name: str, - notebook: "_models.NotebookResource", - if_match: Optional[str] = None, - **kwargs: Any - ) -> AsyncLROPoller["_models.NotebookResource"]: - """Creates or updates a Note Book. - - :param notebook_name: The notebook name. - :type notebook_name: str - :param notebook: Note book resource definition. - :type notebook: ~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookResource - :param if_match: ETag of the Note book entity. Should only be specified for update, for which - it should match existing entity or can be * for unconditional update. - :type if_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either NotebookResource or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookResource] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.NotebookResource"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._create_or_update_notebook_initial( - notebook_name=notebook_name, - notebook=notebook, - if_match=if_match, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('NotebookResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'notebookName': self._serialize.url("notebook_name", notebook_name, 'str'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create_or_update_notebook.metadata = {'url': '/notebooks/{notebookName}'} # type: ignore - - async def get_notebook( - self, - notebook_name: str, - if_none_match: Optional[str] = None, - **kwargs: Any - ) -> Optional["_models.NotebookResource"]: - """Gets a Note Book. - - :param notebook_name: The notebook name. - :type notebook_name: str - :param if_none_match: ETag of the Notebook entity. Should only be specified for get. If the - ETag matches the existing entity tag, or if * was provided, then no content will be returned. - :type if_none_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: NotebookResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookResource or None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.NotebookResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_notebook.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'notebookName': self._serialize.url("notebook_name", notebook_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_none_match is not None: - header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 304]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('NotebookResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_notebook.metadata = {'url': '/notebooks/{notebookName}'} # type: ignore - - async def _delete_notebook_initial( - self, - notebook_name: str, - **kwargs: Any - ) -> None: - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._delete_notebook_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'notebookName': self._serialize.url("notebook_name", notebook_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.delete(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _delete_notebook_initial.metadata = {'url': '/notebooks/{notebookName}'} # type: ignore - - async def begin_delete_notebook( - self, - notebook_name: str, - **kwargs: Any - ) -> AsyncLROPoller[None]: - """Deletes a Note book. - - :param notebook_name: The notebook name. - :type notebook_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._delete_notebook_initial( - notebook_name=notebook_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'notebookName': self._serialize.url("notebook_name", notebook_name, 'str'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_delete_notebook.metadata = {'url': '/notebooks/{notebookName}'} # type: ignore - - async def _rename_notebook_initial( - self, - notebook_name: str, - new_name: Optional[str] = None, - **kwargs: Any - ) -> None: - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _request = _models.ArtifactRenameRequest(new_name=new_name) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._rename_notebook_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'notebookName': self._serialize.url("notebook_name", notebook_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_request, 'ArtifactRenameRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _rename_notebook_initial.metadata = {'url': '/notebooks/{notebookName}/rename'} # type: ignore - - async def begin_rename_notebook( - self, - notebook_name: str, - new_name: Optional[str] = None, - **kwargs: Any - ) -> AsyncLROPoller[None]: - """Renames a notebook. - - :param notebook_name: The notebook name. - :type notebook_name: str - :param new_name: New name of the artifact. - :type new_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._rename_notebook_initial( - notebook_name=notebook_name, - new_name=new_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'notebookName': self._serialize.url("notebook_name", notebook_name, 'str'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_rename_notebook.metadata = {'url': '/notebooks/{notebookName}/rename'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_operation_result_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_operation_result_operations.py deleted file mode 100644 index 6903a0435f95..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_operation_result_operations.py +++ /dev/null @@ -1,92 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, Callable, Dict, Generic, Optional, TypeVar -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class OperationResultOperations: - """OperationResultOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - async def get( - self, - operation_id: str, - **kwargs: Any - ) -> None: - """Get operation result. - - :param operation_id: Operation ID. - :type operation_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'operationId': self._serialize.url("operation_id", operation_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 201, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - get.metadata = {'url': '/operationResults/{operationId}'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_operation_status_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_operation_status_operations.py deleted file mode 100644 index a811865eaa23..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_operation_status_operations.py +++ /dev/null @@ -1,92 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, Callable, Dict, Generic, Optional, TypeVar -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class OperationStatusOperations: - """OperationStatusOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - async def get( - self, - operation_id: str, - **kwargs: Any - ) -> None: - """Get operation status. - - :param operation_id: Operation ID. - :type operation_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'operationId': self._serialize.url("operation_id", operation_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 201, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - get.metadata = {'url': '/operationStatuses/{operationId}'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_pipeline_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_pipeline_operations.py deleted file mode 100644 index db0b75f8d5e0..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_pipeline_operations.py +++ /dev/null @@ -1,606 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, AsyncIterable, Callable, Dict, Generic, Optional, TypeVar, Union -import warnings - -from azure.core.async_paging import AsyncItemPaged, AsyncList -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest -from azure.core.polling import AsyncLROPoller, AsyncNoPolling, AsyncPollingMethod -from azure.core.polling.async_base_polling import AsyncLROBasePolling - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class PipelineOperations: - """PipelineOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get_pipelines_by_workspace( - self, - **kwargs: Any - ) -> AsyncIterable["_models.PipelineListResponse"]: - """Lists pipelines. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either PipelineListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.PipelineListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.get_pipelines_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - async def extract_data(pipeline_response): - deserialized = self._deserialize('PipelineListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, AsyncList(list_of_elem) - - async def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return AsyncItemPaged( - get_next, extract_data - ) - get_pipelines_by_workspace.metadata = {'url': '/pipelines'} # type: ignore - - async def _create_or_update_pipeline_initial( - self, - pipeline_name: str, - pipeline: "_models.PipelineResource", - if_match: Optional[str] = None, - **kwargs: Any - ) -> Optional["_models.PipelineResource"]: - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.PipelineResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._create_or_update_pipeline_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'pipelineName': self._serialize.url("pipeline_name", pipeline_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_match is not None: - header_parameters['If-Match'] = self._serialize.header("if_match", if_match, 'str') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(pipeline, 'PipelineResource') - body_content_kwargs['content'] = body_content - request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('PipelineResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _create_or_update_pipeline_initial.metadata = {'url': '/pipelines/{pipelineName}'} # type: ignore - - async def begin_create_or_update_pipeline( - self, - pipeline_name: str, - pipeline: "_models.PipelineResource", - if_match: Optional[str] = None, - **kwargs: Any - ) -> AsyncLROPoller["_models.PipelineResource"]: - """Creates or updates a pipeline. - - :param pipeline_name: The pipeline name. - :type pipeline_name: str - :param pipeline: Pipeline resource definition. - :type pipeline: ~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineResource - :param if_match: ETag of the pipeline entity. Should only be specified for update, for which - it should match existing entity or can be * for unconditional update. - :type if_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either PipelineResource or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineResource] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.PipelineResource"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._create_or_update_pipeline_initial( - pipeline_name=pipeline_name, - pipeline=pipeline, - if_match=if_match, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('PipelineResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'pipelineName': self._serialize.url("pipeline_name", pipeline_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create_or_update_pipeline.metadata = {'url': '/pipelines/{pipelineName}'} # type: ignore - - async def get_pipeline( - self, - pipeline_name: str, - if_none_match: Optional[str] = None, - **kwargs: Any - ) -> Optional["_models.PipelineResource"]: - """Gets a pipeline. - - :param pipeline_name: The pipeline name. - :type pipeline_name: str - :param if_none_match: ETag of the pipeline entity. Should only be specified for get. If the - ETag matches the existing entity tag, or if * was provided, then no content will be returned. - :type if_none_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: PipelineResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineResource or None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.PipelineResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_pipeline.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'pipelineName': self._serialize.url("pipeline_name", pipeline_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_none_match is not None: - header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 304]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('PipelineResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_pipeline.metadata = {'url': '/pipelines/{pipelineName}'} # type: ignore - - async def _delete_pipeline_initial( - self, - pipeline_name: str, - **kwargs: Any - ) -> None: - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._delete_pipeline_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'pipelineName': self._serialize.url("pipeline_name", pipeline_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.delete(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _delete_pipeline_initial.metadata = {'url': '/pipelines/{pipelineName}'} # type: ignore - - async def begin_delete_pipeline( - self, - pipeline_name: str, - **kwargs: Any - ) -> AsyncLROPoller[None]: - """Deletes a pipeline. - - :param pipeline_name: The pipeline name. - :type pipeline_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._delete_pipeline_initial( - pipeline_name=pipeline_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'pipelineName': self._serialize.url("pipeline_name", pipeline_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_delete_pipeline.metadata = {'url': '/pipelines/{pipelineName}'} # type: ignore - - async def _rename_pipeline_initial( - self, - pipeline_name: str, - new_name: Optional[str] = None, - **kwargs: Any - ) -> None: - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _request = _models.ArtifactRenameRequest(new_name=new_name) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._rename_pipeline_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'pipelineName': self._serialize.url("pipeline_name", pipeline_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_request, 'ArtifactRenameRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _rename_pipeline_initial.metadata = {'url': '/pipelines/{pipelineName}/rename'} # type: ignore - - async def begin_rename_pipeline( - self, - pipeline_name: str, - new_name: Optional[str] = None, - **kwargs: Any - ) -> AsyncLROPoller[None]: - """Renames a pipeline. - - :param pipeline_name: The pipeline name. - :type pipeline_name: str - :param new_name: New name of the artifact. - :type new_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._rename_pipeline_initial( - pipeline_name=pipeline_name, - new_name=new_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'pipelineName': self._serialize.url("pipeline_name", pipeline_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_rename_pipeline.metadata = {'url': '/pipelines/{pipelineName}/rename'} # type: ignore - - async def create_pipeline_run( - self, - pipeline_name: str, - reference_pipeline_run_id: Optional[str] = None, - is_recovery: Optional[bool] = None, - start_activity_name: Optional[str] = None, - parameters: Optional[Dict[str, Any]] = None, - **kwargs: Any - ) -> "_models.CreateRunResponse": - """Creates a run of a pipeline. - - :param pipeline_name: The pipeline name. - :type pipeline_name: str - :param reference_pipeline_run_id: The pipeline run identifier. If run ID is specified the - parameters of the specified run will be used to create a new run. - :type reference_pipeline_run_id: str - :param is_recovery: Recovery mode flag. If recovery mode is set to true, the specified - referenced pipeline run and the new run will be grouped under the same groupId. - :type is_recovery: bool - :param start_activity_name: In recovery mode, the rerun will start from this activity. If not - specified, all activities will run. - :type start_activity_name: str - :param parameters: Parameters of the pipeline run. These parameters will be used only if the - runId is not specified. - :type parameters: dict[str, any] - :keyword callable cls: A custom type or function that will be passed the direct response - :return: CreateRunResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.CreateRunResponse - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.CreateRunResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self.create_pipeline_run.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'pipelineName': self._serialize.url("pipeline_name", pipeline_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - if reference_pipeline_run_id is not None: - query_parameters['referencePipelineRunId'] = self._serialize.query("reference_pipeline_run_id", reference_pipeline_run_id, 'str') - if is_recovery is not None: - query_parameters['isRecovery'] = self._serialize.query("is_recovery", is_recovery, 'bool') - if start_activity_name is not None: - query_parameters['startActivityName'] = self._serialize.query("start_activity_name", start_activity_name, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - if parameters is not None: - body_content = self._serialize.body(parameters, '{object}') - else: - body_content = None - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('CreateRunResponse', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - create_pipeline_run.metadata = {'url': '/pipelines/{pipelineName}/createRun'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_pipeline_run_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_pipeline_run_operations.py deleted file mode 100644 index b82e57beb07e..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_pipeline_run_operations.py +++ /dev/null @@ -1,279 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, Callable, Dict, Generic, Optional, TypeVar -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class PipelineRunOperations: - """PipelineRunOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - async def query_pipeline_runs_by_workspace( - self, - filter_parameters: "_models.RunFilterParameters", - **kwargs: Any - ) -> "_models.PipelineRunsQueryResponse": - """Query pipeline runs in the workspace based on input filter conditions. - - :param filter_parameters: Parameters to filter the pipeline run. - :type filter_parameters: ~azure.synapse.artifacts.v2021_06_01_preview.models.RunFilterParameters - :keyword callable cls: A custom type or function that will be passed the direct response - :return: PipelineRunsQueryResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineRunsQueryResponse - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.PipelineRunsQueryResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self.query_pipeline_runs_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(filter_parameters, 'RunFilterParameters') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('PipelineRunsQueryResponse', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - query_pipeline_runs_by_workspace.metadata = {'url': '/queryPipelineRuns'} # type: ignore - - async def get_pipeline_run( - self, - run_id: str, - **kwargs: Any - ) -> "_models.PipelineRun": - """Get a pipeline run by its run ID. - - :param run_id: The pipeline run identifier. - :type run_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: PipelineRun, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineRun - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.PipelineRun"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_pipeline_run.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'runId': self._serialize.url("run_id", run_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('PipelineRun', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_pipeline_run.metadata = {'url': '/pipelineruns/{runId}'} # type: ignore - - async def query_activity_runs( - self, - pipeline_name: str, - run_id: str, - filter_parameters: "_models.RunFilterParameters", - **kwargs: Any - ) -> "_models.ActivityRunsQueryResponse": - """Query activity runs based on input filter conditions. - - :param pipeline_name: The pipeline name. - :type pipeline_name: str - :param run_id: The pipeline run identifier. - :type run_id: str - :param filter_parameters: Parameters to filter the activity runs. - :type filter_parameters: ~azure.synapse.artifacts.v2021_06_01_preview.models.RunFilterParameters - :keyword callable cls: A custom type or function that will be passed the direct response - :return: ActivityRunsQueryResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityRunsQueryResponse - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.ActivityRunsQueryResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self.query_activity_runs.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'pipelineName': self._serialize.url("pipeline_name", pipeline_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - 'runId': self._serialize.url("run_id", run_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(filter_parameters, 'RunFilterParameters') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('ActivityRunsQueryResponse', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - query_activity_runs.metadata = {'url': '/pipelines/{pipelineName}/pipelineruns/{runId}/queryActivityruns'} # type: ignore - - async def cancel_pipeline_run( - self, - run_id: str, - is_recursive: Optional[bool] = None, - **kwargs: Any - ) -> None: - """Cancel a pipeline run by its run ID. - - :param run_id: The pipeline run identifier. - :type run_id: str - :param is_recursive: If true, cancel all the Child pipelines that are triggered by the current - pipeline. - :type is_recursive: bool - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.cancel_pipeline_run.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'runId': self._serialize.url("run_id", run_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - if is_recursive is not None: - query_parameters['isRecursive'] = self._serialize.query("is_recursive", is_recursive, 'bool') - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - cancel_pipeline_run.metadata = {'url': '/pipelineruns/{runId}/cancel'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_spark_job_definition_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_spark_job_definition_operations.py deleted file mode 100644 index babfaf4270f9..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_spark_job_definition_operations.py +++ /dev/null @@ -1,753 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, AsyncIterable, Callable, Dict, Generic, Optional, TypeVar, Union -import warnings - -from azure.core.async_paging import AsyncItemPaged, AsyncList -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest -from azure.core.polling import AsyncLROPoller, AsyncNoPolling, AsyncPollingMethod -from azure.core.polling.async_base_polling import AsyncLROBasePolling - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class SparkJobDefinitionOperations: - """SparkJobDefinitionOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get_spark_job_definitions_by_workspace( - self, - **kwargs: Any - ) -> AsyncIterable["_models.SparkJobDefinitionsListResponse"]: - """Lists spark job definitions. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either SparkJobDefinitionsListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobDefinitionsListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.SparkJobDefinitionsListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.get_spark_job_definitions_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - async def extract_data(pipeline_response): - deserialized = self._deserialize('SparkJobDefinitionsListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, AsyncList(list_of_elem) - - async def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return AsyncItemPaged( - get_next, extract_data - ) - get_spark_job_definitions_by_workspace.metadata = {'url': '/sparkJobDefinitions'} # type: ignore - - async def _create_or_update_spark_job_definition_initial( - self, - spark_job_definition_name: str, - properties: "_models.SparkJobDefinition", - if_match: Optional[str] = None, - **kwargs: Any - ) -> Optional["_models.SparkJobDefinitionResource"]: - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.SparkJobDefinitionResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _spark_job_definition = _models.SparkJobDefinitionResource(properties=properties) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._create_or_update_spark_job_definition_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sparkJobDefinitionName': self._serialize.url("spark_job_definition_name", spark_job_definition_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_match is not None: - header_parameters['If-Match'] = self._serialize.header("if_match", if_match, 'str') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_spark_job_definition, 'SparkJobDefinitionResource') - body_content_kwargs['content'] = body_content - request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('SparkJobDefinitionResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _create_or_update_spark_job_definition_initial.metadata = {'url': '/sparkJobDefinitions/{sparkJobDefinitionName}'} # type: ignore - - async def begin_create_or_update_spark_job_definition( - self, - spark_job_definition_name: str, - properties: "_models.SparkJobDefinition", - if_match: Optional[str] = None, - **kwargs: Any - ) -> AsyncLROPoller["_models.SparkJobDefinitionResource"]: - """Creates or updates a Spark Job Definition. - - :param spark_job_definition_name: The spark job definition name. - :type spark_job_definition_name: str - :param properties: Properties of spark job definition. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobDefinition - :param if_match: ETag of the Spark Job Definition entity. Should only be specified for update, - for which it should match existing entity or can be * for unconditional update. - :type if_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either SparkJobDefinitionResource or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobDefinitionResource] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.SparkJobDefinitionResource"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._create_or_update_spark_job_definition_initial( - spark_job_definition_name=spark_job_definition_name, - properties=properties, - if_match=if_match, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('SparkJobDefinitionResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sparkJobDefinitionName': self._serialize.url("spark_job_definition_name", spark_job_definition_name, 'str'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create_or_update_spark_job_definition.metadata = {'url': '/sparkJobDefinitions/{sparkJobDefinitionName}'} # type: ignore - - async def get_spark_job_definition( - self, - spark_job_definition_name: str, - if_none_match: Optional[str] = None, - **kwargs: Any - ) -> Optional["_models.SparkJobDefinitionResource"]: - """Gets a Spark Job Definition. - - :param spark_job_definition_name: The spark job definition name. - :type spark_job_definition_name: str - :param if_none_match: ETag of the Spark Job Definition entity. Should only be specified for - get. If the ETag matches the existing entity tag, or if * was provided, then no content will be - returned. - :type if_none_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: SparkJobDefinitionResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobDefinitionResource or None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.SparkJobDefinitionResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_spark_job_definition.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sparkJobDefinitionName': self._serialize.url("spark_job_definition_name", spark_job_definition_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_none_match is not None: - header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 304]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('SparkJobDefinitionResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_spark_job_definition.metadata = {'url': '/sparkJobDefinitions/{sparkJobDefinitionName}'} # type: ignore - - async def _delete_spark_job_definition_initial( - self, - spark_job_definition_name: str, - **kwargs: Any - ) -> None: - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._delete_spark_job_definition_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sparkJobDefinitionName': self._serialize.url("spark_job_definition_name", spark_job_definition_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.delete(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _delete_spark_job_definition_initial.metadata = {'url': '/sparkJobDefinitions/{sparkJobDefinitionName}'} # type: ignore - - async def begin_delete_spark_job_definition( - self, - spark_job_definition_name: str, - **kwargs: Any - ) -> AsyncLROPoller[None]: - """Deletes a Spark Job Definition. - - :param spark_job_definition_name: The spark job definition name. - :type spark_job_definition_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._delete_spark_job_definition_initial( - spark_job_definition_name=spark_job_definition_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sparkJobDefinitionName': self._serialize.url("spark_job_definition_name", spark_job_definition_name, 'str'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_delete_spark_job_definition.metadata = {'url': '/sparkJobDefinitions/{sparkJobDefinitionName}'} # type: ignore - - async def _execute_spark_job_definition_initial( - self, - spark_job_definition_name: str, - **kwargs: Any - ) -> "_models.SparkBatchJob": - cls = kwargs.pop('cls', None) # type: ClsType["_models.SparkBatchJob"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._execute_spark_job_definition_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sparkJobDefinitionName': self._serialize.url("spark_job_definition_name", spark_job_definition_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if response.status_code == 200: - deserialized = self._deserialize('SparkBatchJob', pipeline_response) - - if response.status_code == 202: - deserialized = self._deserialize('SparkBatchJob', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _execute_spark_job_definition_initial.metadata = {'url': '/sparkJobDefinitions/{sparkJobDefinitionName}/execute'} # type: ignore - - async def begin_execute_spark_job_definition( - self, - spark_job_definition_name: str, - **kwargs: Any - ) -> AsyncLROPoller["_models.SparkBatchJob"]: - """Executes the spark job definition. - - :param spark_job_definition_name: The spark job definition name. - :type spark_job_definition_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either SparkBatchJob or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.SparkBatchJob] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.SparkBatchJob"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._execute_spark_job_definition_initial( - spark_job_definition_name=spark_job_definition_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('SparkBatchJob', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sparkJobDefinitionName': self._serialize.url("spark_job_definition_name", spark_job_definition_name, 'str'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, lro_options={'final-state-via': 'location'}, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_execute_spark_job_definition.metadata = {'url': '/sparkJobDefinitions/{sparkJobDefinitionName}/execute'} # type: ignore - - async def _rename_spark_job_definition_initial( - self, - spark_job_definition_name: str, - new_name: Optional[str] = None, - **kwargs: Any - ) -> None: - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _request = _models.ArtifactRenameRequest(new_name=new_name) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._rename_spark_job_definition_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sparkJobDefinitionName': self._serialize.url("spark_job_definition_name", spark_job_definition_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_request, 'ArtifactRenameRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _rename_spark_job_definition_initial.metadata = {'url': '/sparkJobDefinitions/{sparkJobDefinitionName}/rename'} # type: ignore - - async def begin_rename_spark_job_definition( - self, - spark_job_definition_name: str, - new_name: Optional[str] = None, - **kwargs: Any - ) -> AsyncLROPoller[None]: - """Renames a sparkJobDefinition. - - :param spark_job_definition_name: The spark job definition name. - :type spark_job_definition_name: str - :param new_name: New name of the artifact. - :type new_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._rename_spark_job_definition_initial( - spark_job_definition_name=spark_job_definition_name, - new_name=new_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sparkJobDefinitionName': self._serialize.url("spark_job_definition_name", spark_job_definition_name, 'str'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_rename_spark_job_definition.metadata = {'url': '/sparkJobDefinitions/{sparkJobDefinitionName}/rename'} # type: ignore - - async def _debug_spark_job_definition_initial( - self, - properties: "_models.SparkJobDefinition", - **kwargs: Any - ) -> "_models.SparkBatchJob": - cls = kwargs.pop('cls', None) # type: ClsType["_models.SparkBatchJob"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _spark_job_definition_azure_resource = _models.SparkJobDefinitionResource(properties=properties) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._debug_spark_job_definition_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_spark_job_definition_azure_resource, 'SparkJobDefinitionResource') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if response.status_code == 200: - deserialized = self._deserialize('SparkBatchJob', pipeline_response) - - if response.status_code == 202: - deserialized = self._deserialize('SparkBatchJob', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _debug_spark_job_definition_initial.metadata = {'url': '/debugSparkJobDefinition'} # type: ignore - - async def begin_debug_spark_job_definition( - self, - properties: "_models.SparkJobDefinition", - **kwargs: Any - ) -> AsyncLROPoller["_models.SparkBatchJob"]: - """Debug the spark job definition. - - :param properties: Properties of spark job definition. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobDefinition - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either SparkBatchJob or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.SparkBatchJob] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.SparkBatchJob"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._debug_spark_job_definition_initial( - properties=properties, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('SparkBatchJob', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, lro_options={'final-state-via': 'location'}, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_debug_spark_job_definition.metadata = {'url': '/debugSparkJobDefinition'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_sql_pools_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_sql_pools_operations.py deleted file mode 100644 index f5b4a16d331d..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_sql_pools_operations.py +++ /dev/null @@ -1,146 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, Callable, Dict, Generic, Optional, TypeVar -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class SqlPoolsOperations: - """SqlPoolsOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - async def list( - self, - **kwargs: Any - ) -> "_models.SqlPoolInfoListResult": - """List Sql Pools. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolInfoListResult, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPoolInfoListResult - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.SqlPoolInfoListResult"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.list.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('SqlPoolInfoListResult', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - list.metadata = {'url': '/sqlPools'} # type: ignore - - async def get( - self, - sql_pool_name: str, - **kwargs: Any - ) -> "_models.SqlPool": - """Get Sql Pool. - - :param sql_pool_name: The Sql Pool name. - :type sql_pool_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPool, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPool - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.SqlPool"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sqlPoolName': self._serialize.url("sql_pool_name", sql_pool_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('SqlPool', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get.metadata = {'url': '/sqlPools/{sqlPoolName}'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_sql_script_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_sql_script_operations.py deleted file mode 100644 index c493f8a49049..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_sql_script_operations.py +++ /dev/null @@ -1,521 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, AsyncIterable, Callable, Dict, Generic, Optional, TypeVar, Union -import warnings - -from azure.core.async_paging import AsyncItemPaged, AsyncList -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest -from azure.core.polling import AsyncLROPoller, AsyncNoPolling, AsyncPollingMethod -from azure.core.polling.async_base_polling import AsyncLROBasePolling - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class SqlScriptOperations: - """SqlScriptOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get_sql_scripts_by_workspace( - self, - **kwargs: Any - ) -> AsyncIterable["_models.SqlScriptsListResponse"]: - """Lists sql scripts. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either SqlScriptsListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.SqlScriptsListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.SqlScriptsListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.get_sql_scripts_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - async def extract_data(pipeline_response): - deserialized = self._deserialize('SqlScriptsListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, AsyncList(list_of_elem) - - async def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return AsyncItemPaged( - get_next, extract_data - ) - get_sql_scripts_by_workspace.metadata = {'url': '/sqlScripts'} # type: ignore - - async def _create_or_update_sql_script_initial( - self, - sql_script_name: str, - sql_script: "_models.SqlScriptResource", - if_match: Optional[str] = None, - **kwargs: Any - ) -> Optional["_models.SqlScriptResource"]: - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.SqlScriptResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._create_or_update_sql_script_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sqlScriptName': self._serialize.url("sql_script_name", sql_script_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_match is not None: - header_parameters['If-Match'] = self._serialize.header("if_match", if_match, 'str') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(sql_script, 'SqlScriptResource') - body_content_kwargs['content'] = body_content - request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('SqlScriptResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _create_or_update_sql_script_initial.metadata = {'url': '/sqlScripts/{sqlScriptName}'} # type: ignore - - async def begin_create_or_update_sql_script( - self, - sql_script_name: str, - sql_script: "_models.SqlScriptResource", - if_match: Optional[str] = None, - **kwargs: Any - ) -> AsyncLROPoller["_models.SqlScriptResource"]: - """Creates or updates a Sql Script. - - :param sql_script_name: The sql script name. - :type sql_script_name: str - :param sql_script: Sql Script resource definition. - :type sql_script: ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlScriptResource - :param if_match: ETag of the SQL script entity. Should only be specified for update, for which - it should match existing entity or can be * for unconditional update. - :type if_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either SqlScriptResource or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.SqlScriptResource] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.SqlScriptResource"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._create_or_update_sql_script_initial( - sql_script_name=sql_script_name, - sql_script=sql_script, - if_match=if_match, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('SqlScriptResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sqlScriptName': self._serialize.url("sql_script_name", sql_script_name, 'str'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create_or_update_sql_script.metadata = {'url': '/sqlScripts/{sqlScriptName}'} # type: ignore - - async def get_sql_script( - self, - sql_script_name: str, - if_none_match: Optional[str] = None, - **kwargs: Any - ) -> Optional["_models.SqlScriptResource"]: - """Gets a sql script. - - :param sql_script_name: The sql script name. - :type sql_script_name: str - :param if_none_match: ETag of the sql compute entity. Should only be specified for get. If the - ETag matches the existing entity tag, or if * was provided, then no content will be returned. - :type if_none_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlScriptResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlScriptResource or None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.SqlScriptResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_sql_script.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sqlScriptName': self._serialize.url("sql_script_name", sql_script_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_none_match is not None: - header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 304]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('SqlScriptResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_sql_script.metadata = {'url': '/sqlScripts/{sqlScriptName}'} # type: ignore - - async def _delete_sql_script_initial( - self, - sql_script_name: str, - **kwargs: Any - ) -> None: - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._delete_sql_script_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sqlScriptName': self._serialize.url("sql_script_name", sql_script_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.delete(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _delete_sql_script_initial.metadata = {'url': '/sqlScripts/{sqlScriptName}'} # type: ignore - - async def begin_delete_sql_script( - self, - sql_script_name: str, - **kwargs: Any - ) -> AsyncLROPoller[None]: - """Deletes a Sql Script. - - :param sql_script_name: The sql script name. - :type sql_script_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._delete_sql_script_initial( - sql_script_name=sql_script_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sqlScriptName': self._serialize.url("sql_script_name", sql_script_name, 'str'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_delete_sql_script.metadata = {'url': '/sqlScripts/{sqlScriptName}'} # type: ignore - - async def _rename_sql_script_initial( - self, - sql_script_name: str, - new_name: Optional[str] = None, - **kwargs: Any - ) -> None: - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _request = _models.ArtifactRenameRequest(new_name=new_name) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._rename_sql_script_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sqlScriptName': self._serialize.url("sql_script_name", sql_script_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_request, 'ArtifactRenameRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _rename_sql_script_initial.metadata = {'url': '/sqlScripts/{sqlScriptName}/rename'} # type: ignore - - async def begin_rename_sql_script( - self, - sql_script_name: str, - new_name: Optional[str] = None, - **kwargs: Any - ) -> AsyncLROPoller[None]: - """Renames a sqlScript. - - :param sql_script_name: The sql script name. - :type sql_script_name: str - :param new_name: New name of the artifact. - :type new_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._rename_sql_script_initial( - sql_script_name=sql_script_name, - new_name=new_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sqlScriptName': self._serialize.url("sql_script_name", sql_script_name, 'str'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_rename_sql_script.metadata = {'url': '/sqlScripts/{sqlScriptName}/rename'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_trigger_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_trigger_operations.py deleted file mode 100644 index d55686413ef4..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_trigger_operations.py +++ /dev/null @@ -1,888 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, AsyncIterable, Callable, Dict, Generic, Optional, TypeVar, Union -import warnings - -from azure.core.async_paging import AsyncItemPaged, AsyncList -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest -from azure.core.polling import AsyncLROPoller, AsyncNoPolling, AsyncPollingMethod -from azure.core.polling.async_base_polling import AsyncLROBasePolling - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class TriggerOperations: - """TriggerOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get_triggers_by_workspace( - self, - **kwargs: Any - ) -> AsyncIterable["_models.TriggerListResponse"]: - """Lists triggers. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either TriggerListResponse or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.TriggerListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.get_triggers_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - async def extract_data(pipeline_response): - deserialized = self._deserialize('TriggerListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, AsyncList(list_of_elem) - - async def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return AsyncItemPaged( - get_next, extract_data - ) - get_triggers_by_workspace.metadata = {'url': '/triggers'} # type: ignore - - async def _create_or_update_trigger_initial( - self, - trigger_name: str, - properties: "_models.Trigger", - if_match: Optional[str] = None, - **kwargs: Any - ) -> Optional["_models.TriggerResource"]: - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.TriggerResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _trigger = _models.TriggerResource(properties=properties) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._create_or_update_trigger_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_match is not None: - header_parameters['If-Match'] = self._serialize.header("if_match", if_match, 'str') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_trigger, 'TriggerResource') - body_content_kwargs['content'] = body_content - request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('TriggerResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _create_or_update_trigger_initial.metadata = {'url': '/triggers/{triggerName}'} # type: ignore - - async def begin_create_or_update_trigger( - self, - trigger_name: str, - properties: "_models.Trigger", - if_match: Optional[str] = None, - **kwargs: Any - ) -> AsyncLROPoller["_models.TriggerResource"]: - """Creates or updates a trigger. - - :param trigger_name: The trigger name. - :type trigger_name: str - :param properties: Properties of the trigger. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.Trigger - :param if_match: ETag of the trigger entity. Should only be specified for update, for which it - should match existing entity or can be * for unconditional update. - :type if_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either TriggerResource or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerResource] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.TriggerResource"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._create_or_update_trigger_initial( - trigger_name=trigger_name, - properties=properties, - if_match=if_match, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('TriggerResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create_or_update_trigger.metadata = {'url': '/triggers/{triggerName}'} # type: ignore - - async def get_trigger( - self, - trigger_name: str, - if_none_match: Optional[str] = None, - **kwargs: Any - ) -> Optional["_models.TriggerResource"]: - """Gets a trigger. - - :param trigger_name: The trigger name. - :type trigger_name: str - :param if_none_match: ETag of the trigger entity. Should only be specified for get. If the ETag - matches the existing entity tag, or if * was provided, then no content will be returned. - :type if_none_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: TriggerResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerResource or None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.TriggerResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_trigger.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_none_match is not None: - header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 304]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('TriggerResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_trigger.metadata = {'url': '/triggers/{triggerName}'} # type: ignore - - async def _delete_trigger_initial( - self, - trigger_name: str, - **kwargs: Any - ) -> None: - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._delete_trigger_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.delete(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _delete_trigger_initial.metadata = {'url': '/triggers/{triggerName}'} # type: ignore - - async def begin_delete_trigger( - self, - trigger_name: str, - **kwargs: Any - ) -> AsyncLROPoller[None]: - """Deletes a trigger. - - :param trigger_name: The trigger name. - :type trigger_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._delete_trigger_initial( - trigger_name=trigger_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_delete_trigger.metadata = {'url': '/triggers/{triggerName}'} # type: ignore - - async def _subscribe_trigger_to_events_initial( - self, - trigger_name: str, - **kwargs: Any - ) -> Optional["_models.TriggerSubscriptionOperationStatus"]: - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.TriggerSubscriptionOperationStatus"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._subscribe_trigger_to_events_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('TriggerSubscriptionOperationStatus', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _subscribe_trigger_to_events_initial.metadata = {'url': '/triggers/{triggerName}/subscribeToEvents'} # type: ignore - - async def begin_subscribe_trigger_to_events( - self, - trigger_name: str, - **kwargs: Any - ) -> AsyncLROPoller["_models.TriggerSubscriptionOperationStatus"]: - """Subscribe event trigger to events. - - :param trigger_name: The trigger name. - :type trigger_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either TriggerSubscriptionOperationStatus or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerSubscriptionOperationStatus] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.TriggerSubscriptionOperationStatus"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._subscribe_trigger_to_events_initial( - trigger_name=trigger_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('TriggerSubscriptionOperationStatus', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_subscribe_trigger_to_events.metadata = {'url': '/triggers/{triggerName}/subscribeToEvents'} # type: ignore - - async def get_event_subscription_status( - self, - trigger_name: str, - **kwargs: Any - ) -> "_models.TriggerSubscriptionOperationStatus": - """Get a trigger's event subscription status. - - :param trigger_name: The trigger name. - :type trigger_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: TriggerSubscriptionOperationStatus, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerSubscriptionOperationStatus - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.TriggerSubscriptionOperationStatus"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_event_subscription_status.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('TriggerSubscriptionOperationStatus', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_event_subscription_status.metadata = {'url': '/triggers/{triggerName}/getEventSubscriptionStatus'} # type: ignore - - async def _unsubscribe_trigger_from_events_initial( - self, - trigger_name: str, - **kwargs: Any - ) -> Optional["_models.TriggerSubscriptionOperationStatus"]: - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.TriggerSubscriptionOperationStatus"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._unsubscribe_trigger_from_events_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('TriggerSubscriptionOperationStatus', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _unsubscribe_trigger_from_events_initial.metadata = {'url': '/triggers/{triggerName}/unsubscribeFromEvents'} # type: ignore - - async def begin_unsubscribe_trigger_from_events( - self, - trigger_name: str, - **kwargs: Any - ) -> AsyncLROPoller["_models.TriggerSubscriptionOperationStatus"]: - """Unsubscribe event trigger from events. - - :param trigger_name: The trigger name. - :type trigger_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either TriggerSubscriptionOperationStatus or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerSubscriptionOperationStatus] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.TriggerSubscriptionOperationStatus"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._unsubscribe_trigger_from_events_initial( - trigger_name=trigger_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('TriggerSubscriptionOperationStatus', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_unsubscribe_trigger_from_events.metadata = {'url': '/triggers/{triggerName}/unsubscribeFromEvents'} # type: ignore - - async def _start_trigger_initial( - self, - trigger_name: str, - **kwargs: Any - ) -> None: - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._start_trigger_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _start_trigger_initial.metadata = {'url': '/triggers/{triggerName}/start'} # type: ignore - - async def begin_start_trigger( - self, - trigger_name: str, - **kwargs: Any - ) -> AsyncLROPoller[None]: - """Starts a trigger. - - :param trigger_name: The trigger name. - :type trigger_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._start_trigger_initial( - trigger_name=trigger_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_start_trigger.metadata = {'url': '/triggers/{triggerName}/start'} # type: ignore - - async def _stop_trigger_initial( - self, - trigger_name: str, - **kwargs: Any - ) -> None: - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._stop_trigger_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _stop_trigger_initial.metadata = {'url': '/triggers/{triggerName}/stop'} # type: ignore - - async def begin_stop_trigger( - self, - trigger_name: str, - **kwargs: Any - ) -> AsyncLROPoller[None]: - """Stops a trigger. - - :param trigger_name: The trigger name. - :type trigger_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be AsyncLROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.AsyncPollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.AsyncLROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, AsyncPollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = await self._stop_trigger_initial( - trigger_name=trigger_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = AsyncLROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = AsyncNoPolling() - else: polling_method = polling - if cont_token: - return AsyncLROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return AsyncLROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_stop_trigger.metadata = {'url': '/triggers/{triggerName}/stop'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_trigger_run_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_trigger_run_operations.py deleted file mode 100644 index d8173b56932a..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_trigger_run_operations.py +++ /dev/null @@ -1,211 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, Callable, Dict, Generic, Optional, TypeVar -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class TriggerRunOperations: - """TriggerRunOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - async def rerun_trigger_instance( - self, - trigger_name: str, - run_id: str, - **kwargs: Any - ) -> None: - """Rerun single trigger instance by runId. - - :param trigger_name: The trigger name. - :type trigger_name: str - :param run_id: The pipeline run identifier. - :type run_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.rerun_trigger_instance.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - 'runId': self._serialize.url("run_id", run_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - rerun_trigger_instance.metadata = {'url': '/triggers/{triggerName}/triggerRuns/{runId}/rerun'} # type: ignore - - async def cancel_trigger_instance( - self, - trigger_name: str, - run_id: str, - **kwargs: Any - ) -> None: - """Cancel single trigger instance by runId. - - :param trigger_name: The trigger name. - :type trigger_name: str - :param run_id: The pipeline run identifier. - :type run_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.cancel_trigger_instance.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - 'runId': self._serialize.url("run_id", run_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - cancel_trigger_instance.metadata = {'url': '/triggers/{triggerName}/triggerRuns/{runId}/cancel'} # type: ignore - - async def query_trigger_runs_by_workspace( - self, - filter_parameters: "_models.RunFilterParameters", - **kwargs: Any - ) -> "_models.TriggerRunsQueryResponse": - """Query trigger runs. - - :param filter_parameters: Parameters to filter the pipeline run. - :type filter_parameters: ~azure.synapse.artifacts.v2021_06_01_preview.models.RunFilterParameters - :keyword callable cls: A custom type or function that will be passed the direct response - :return: TriggerRunsQueryResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRunsQueryResponse - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.TriggerRunsQueryResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self.query_trigger_runs_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(filter_parameters, 'RunFilterParameters') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('TriggerRunsQueryResponse', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - query_trigger_runs_by_workspace.metadata = {'url': '/queryTriggerRuns'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_workspace_git_repo_management_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_workspace_git_repo_management_operations.py deleted file mode 100644 index 5e025b6f2b53..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_workspace_git_repo_management_operations.py +++ /dev/null @@ -1,104 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, Callable, Dict, Generic, Optional, TypeVar -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class WorkspaceGitRepoManagementOperations: - """WorkspaceGitRepoManagementOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - async def get_git_hub_access_token( - self, - git_hub_access_token_request: "_models.GitHubAccessTokenRequest", - client_request_id: Optional[str] = None, - **kwargs: Any - ) -> "_models.GitHubAccessTokenResponse": - """Get the GitHub access token. - - :param git_hub_access_token_request: - :type git_hub_access_token_request: ~azure.synapse.artifacts.v2021_06_01_preview.models.GitHubAccessTokenRequest - :param client_request_id: Can provide a guid, which is helpful for debugging and to provide - better customer support. - :type client_request_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: GitHubAccessTokenResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.GitHubAccessTokenResponse - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.GitHubAccessTokenResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self.get_git_hub_access_token.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if client_request_id is not None: - header_parameters['x-ms-client-request-id'] = self._serialize.header("client_request_id", client_request_id, 'str') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(git_hub_access_token_request, 'GitHubAccessTokenRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response) - - deserialized = self._deserialize('GitHubAccessTokenResponse', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_git_hub_access_token.metadata = {'url': '/getGitHubAccessToken'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_workspace_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_workspace_operations.py deleted file mode 100644 index d8c875a84e75..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/aio/operations/_workspace_operations.py +++ /dev/null @@ -1,91 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import Any, Callable, Dict, Generic, Optional, TypeVar -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import AsyncHttpResponse, HttpRequest - -from ... import models as _models - -T = TypeVar('T') -ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, Dict[str, Any]], Any]] - -class WorkspaceOperations: - """WorkspaceOperations async operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer) -> None: - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - async def get( - self, - **kwargs: Any - ) -> "_models.Workspace": - """Get Workspace. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: Workspace, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.Workspace - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.Workspace"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('Workspace', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get.metadata = {'url': '/workspace'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/models/_artifacts_client_enums.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/models/_artifacts_client_enums.py deleted file mode 100644 index 5259f055fbc7..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/models/_artifacts_client_enums.py +++ /dev/null @@ -1,848 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from enum import Enum, EnumMeta -from six import with_metaclass - -class _CaseInsensitiveEnumMeta(EnumMeta): - def __getitem__(self, name): - return super().__getitem__(name.upper()) - - def __getattr__(cls, name): - """Return the enum member matching `name` - We use __getattr__ instead of descriptors or inserting into the enum - class' __dict__ in order to support `name` and `value` being both - properties for enum members (which live in the class' __dict__) and - enum members themselves. - """ - try: - return cls._member_map_[name.upper()] - except KeyError: - raise AttributeError(name) - - -class AvroCompressionCodec(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - - NONE = "none" - DEFLATE = "deflate" - SNAPPY = "snappy" - XZ = "xz" - BZIP2 = "bzip2" - -class AzureFunctionActivityMethod(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The list of HTTP methods supported by a AzureFunctionActivity. - """ - - GET = "GET" - POST = "POST" - PUT = "PUT" - DELETE = "DELETE" - OPTIONS = "OPTIONS" - HEAD = "HEAD" - TRACE = "TRACE" - -class AzureSearchIndexWriteBehaviorType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Specify the write behavior when upserting documents into Azure Search Index. - """ - - MERGE = "Merge" - UPLOAD = "Upload" - -class BigDataPoolReferenceType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Big data pool reference type. - """ - - BIG_DATA_POOL_REFERENCE = "BigDataPoolReference" - -class BlobEventType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - - MICROSOFT_STORAGE_BLOB_CREATED = "Microsoft.Storage.BlobCreated" - MICROSOFT_STORAGE_BLOB_DELETED = "Microsoft.Storage.BlobDeleted" - -class CassandraSourceReadConsistencyLevels(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The consistency level specifies how many Cassandra servers must respond to a read request - before returning data to the client application. Cassandra checks the specified number of - Cassandra servers for data to satisfy the read request. Must be one of - cassandraSourceReadConsistencyLevels. The default value is 'ONE'. It is case-insensitive. - """ - - ALL = "ALL" - EACH_QUORUM = "EACH_QUORUM" - QUORUM = "QUORUM" - LOCAL_QUORUM = "LOCAL_QUORUM" - ONE = "ONE" - TWO = "TWO" - THREE = "THREE" - LOCAL_ONE = "LOCAL_ONE" - SERIAL = "SERIAL" - LOCAL_SERIAL = "LOCAL_SERIAL" - -class CellOutputType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Execution, display, or stream outputs. - """ - - EXECUTE_RESULT = "execute_result" - DISPLAY_DATA = "display_data" - STREAM = "stream" - ERROR = "error" - -class CompressionCodec(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - - BZIP2 = "bzip2" - GZIP = "gzip" - DEFLATE = "deflate" - ZIP_DEFLATE = "zipDeflate" - SNAPPY = "snappy" - LZ4 = "lz4" - TAR = "tar" - TAR_G_ZIP = "tarGZip" - -class CopyBehaviorType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """All available types of copy behavior. - """ - - PRESERVE_HIERARCHY = "PreserveHierarchy" - FLATTEN_HIERARCHY = "FlattenHierarchy" - MERGE_FILES = "MergeFiles" - -class DataFlowComputeType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Compute type of the cluster which will execute data flow job. - """ - - GENERAL = "General" - MEMORY_OPTIMIZED = "MemoryOptimized" - COMPUTE_OPTIMIZED = "ComputeOptimized" - -class DataFlowReferenceType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Data flow reference type. - """ - - DATA_FLOW_REFERENCE = "DataFlowReference" - -class DatasetCompressionLevel(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - - OPTIMAL = "Optimal" - FASTEST = "Fastest" - -class DatasetReferenceType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Dataset reference type. - """ - - DATASET_REFERENCE = "DatasetReference" - -class DayOfWeek(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - - SUNDAY = "Sunday" - MONDAY = "Monday" - TUESDAY = "Tuesday" - WEDNESDAY = "Wednesday" - THURSDAY = "Thursday" - FRIDAY = "Friday" - SATURDAY = "Saturday" - -class Db2AuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """AuthenticationType to be used for connection. It is mutually exclusive with connectionString - property. - """ - - BASIC = "Basic" - -class DependencyCondition(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - - SUCCEEDED = "Succeeded" - FAILED = "Failed" - SKIPPED = "Skipped" - COMPLETED = "Completed" - -class DynamicsAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The authentication type to connect to Dynamics server. 'Office365' for online scenario, 'Ifd' - for on-premises with Ifd scenario, 'AADServicePrincipal' for Server-To-Server authentication in - online scenario. Type: string (or Expression with resultType string). - """ - - OFFICE365 = "Office365" - IFD = "Ifd" - AAD_SERVICE_PRINCIPAL = "AADServicePrincipal" - -class DynamicsDeploymentType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The deployment type of the Dynamics instance. 'Online' for Dynamics Online and - 'OnPremisesWithIfd' for Dynamics on-premises with Ifd. Type: string (or Expression with - resultType string). - """ - - ONLINE = "Online" - ON_PREMISES_WITH_IFD = "OnPremisesWithIfd" - -class DynamicsServicePrincipalCredentialType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The service principal credential type to use in Server-To-Server authentication. - 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' for certificate. Type: string (or - Expression with resultType string). - """ - - SERVICE_PRINCIPAL_KEY = "ServicePrincipalKey" - SERVICE_PRINCIPAL_CERT = "ServicePrincipalCert" - -class DynamicsSinkWriteBehavior(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Defines values for DynamicsSinkWriteBehavior. - """ - - UPSERT = "Upsert" - -class EventSubscriptionStatus(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Event Subscription Status. - """ - - ENABLED = "Enabled" - PROVISIONING = "Provisioning" - DEPROVISIONING = "Deprovisioning" - DISABLED = "Disabled" - UNKNOWN = "Unknown" - -class ExpressionType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Expression type. - """ - - EXPRESSION = "Expression" - -class FtpAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The authentication type to be used to connect to the FTP server. - """ - - BASIC = "Basic" - ANONYMOUS = "Anonymous" - -class GoogleAdWordsAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The OAuth 2.0 authentication mechanism used for authentication. ServiceAuthentication can only - be used on self-hosted IR. - """ - - SERVICE_AUTHENTICATION = "ServiceAuthentication" - USER_AUTHENTICATION = "UserAuthentication" - -class GoogleBigQueryAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The OAuth 2.0 authentication mechanism used for authentication. ServiceAuthentication can only - be used on self-hosted IR. - """ - - SERVICE_AUTHENTICATION = "ServiceAuthentication" - USER_AUTHENTICATION = "UserAuthentication" - -class HBaseAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The authentication mechanism to use to connect to the HBase server. - """ - - ANONYMOUS = "Anonymous" - BASIC = "Basic" - -class HdiNodeTypes(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The node types on which the script action should be executed. - """ - - HEADNODE = "Headnode" - WORKERNODE = "Workernode" - ZOOKEEPER = "Zookeeper" - -class HDInsightActivityDebugInfoOption(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The HDInsightActivityDebugInfoOption settings to use. - """ - - NONE = "None" - ALWAYS = "Always" - FAILURE = "Failure" - -class HiveAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The authentication method used to access the Hive server. - """ - - ANONYMOUS = "Anonymous" - USERNAME = "Username" - USERNAME_AND_PASSWORD = "UsernameAndPassword" - WINDOWS_AZURE_HD_INSIGHT_SERVICE = "WindowsAzureHDInsightService" - -class HiveServerType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The type of Hive server. - """ - - HIVE_SERVER1 = "HiveServer1" - HIVE_SERVER2 = "HiveServer2" - HIVE_THRIFT_SERVER = "HiveThriftServer" - -class HiveThriftTransportProtocol(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The transport protocol to use in the Thrift layer. - """ - - BINARY = "Binary" - SASL = "SASL" - HTTP = "HTTP " - -class HttpAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The authentication type to be used to connect to the HTTP server. - """ - - BASIC = "Basic" - ANONYMOUS = "Anonymous" - DIGEST = "Digest" - WINDOWS = "Windows" - CLIENT_CERTIFICATE = "ClientCertificate" - -class ImpalaAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The authentication type to use. - """ - - ANONYMOUS = "Anonymous" - SASL_USERNAME = "SASLUsername" - USERNAME_AND_PASSWORD = "UsernameAndPassword" - -class IntegrationRuntimeEdition(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The edition for the SSIS Integration Runtime - """ - - STANDARD = "Standard" - ENTERPRISE = "Enterprise" - -class IntegrationRuntimeEntityReferenceType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The type of this referenced entity. - """ - - INTEGRATION_RUNTIME_REFERENCE = "IntegrationRuntimeReference" - LINKED_SERVICE_REFERENCE = "LinkedServiceReference" - -class IntegrationRuntimeLicenseType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """License type for bringing your own license scenario. - """ - - BASE_PRICE = "BasePrice" - LICENSE_INCLUDED = "LicenseIncluded" - -class IntegrationRuntimeReferenceType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Type of integration runtime. - """ - - INTEGRATION_RUNTIME_REFERENCE = "IntegrationRuntimeReference" - -class IntegrationRuntimeSsisCatalogPricingTier(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The pricing tier for the catalog database. The valid values could be found in - https://azure.microsoft.com/en-us/pricing/details/sql-database/ - """ - - BASIC = "Basic" - STANDARD = "Standard" - PREMIUM = "Premium" - PREMIUM_RS = "PremiumRS" - -class IntegrationRuntimeState(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The state of integration runtime. - """ - - INITIAL = "Initial" - STOPPED = "Stopped" - STARTED = "Started" - STARTING = "Starting" - STOPPING = "Stopping" - NEED_REGISTRATION = "NeedRegistration" - ONLINE = "Online" - LIMITED = "Limited" - OFFLINE = "Offline" - ACCESS_DENIED = "AccessDenied" - -class IntegrationRuntimeType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The type of integration runtime. - """ - - MANAGED = "Managed" - SELF_HOSTED = "SelfHosted" - -class JsonFormatFilePattern(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """JSON format file pattern. A property of JsonFormat. - """ - - SET_OF_OBJECTS = "setOfObjects" - ARRAY_OF_OBJECTS = "arrayOfObjects" - -class JsonWriteFilePattern(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """File pattern of JSON. This setting controls the way a collection of JSON objects will be - treated. The default value is 'setOfObjects'. It is case-sensitive. - """ - - SET_OF_OBJECTS = "setOfObjects" - ARRAY_OF_OBJECTS = "arrayOfObjects" - -class MongoDbAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The authentication type to be used to connect to the MongoDB database. - """ - - BASIC = "Basic" - ANONYMOUS = "Anonymous" - -class NetezzaPartitionOption(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The partition mechanism that will be used for Netezza read in parallel. - """ - - NONE = "None" - DATA_SLICE = "DataSlice" - DYNAMIC_RANGE = "DynamicRange" - -class NodeSize(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The level of compute power that each node in the Big Data pool has. - """ - - NONE = "None" - SMALL = "Small" - MEDIUM = "Medium" - LARGE = "Large" - X_LARGE = "XLarge" - XX_LARGE = "XXLarge" - XXX_LARGE = "XXXLarge" - -class NodeSizeFamily(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The kind of nodes that the Big Data pool provides. - """ - - NONE = "None" - MEMORY_OPTIMIZED = "MemoryOptimized" - -class NotebookReferenceType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Synapse notebook reference type. - """ - - NOTEBOOK_REFERENCE = "NotebookReference" - -class ODataAadServicePrincipalCredentialType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Specify the credential type (key or cert) is used for service principal. - """ - - SERVICE_PRINCIPAL_KEY = "ServicePrincipalKey" - SERVICE_PRINCIPAL_CERT = "ServicePrincipalCert" - -class ODataAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Type of authentication used to connect to the OData service. - """ - - BASIC = "Basic" - ANONYMOUS = "Anonymous" - WINDOWS = "Windows" - AAD_SERVICE_PRINCIPAL = "AadServicePrincipal" - MANAGED_SERVICE_IDENTITY = "ManagedServiceIdentity" - -class OraclePartitionOption(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The partition mechanism that will be used for Oracle read in parallel. - """ - - NONE = "None" - PHYSICAL_PARTITIONS_OF_TABLE = "PhysicalPartitionsOfTable" - DYNAMIC_RANGE = "DynamicRange" - -class OrcCompressionCodec(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - - NONE = "none" - ZLIB = "zlib" - SNAPPY = "snappy" - LZO = "lzo" - -class ParameterType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Parameter type. - """ - - OBJECT = "Object" - STRING = "String" - INT = "Int" - FLOAT = "Float" - BOOL = "Bool" - ARRAY = "Array" - SECURE_STRING = "SecureString" - -class ParquetCompressionCodecEnum(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - - NONE = "none" - GZIP = "gzip" - SNAPPY = "snappy" - LZO = "lzo" - -class PhoenixAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The authentication mechanism used to connect to the Phoenix server. - """ - - ANONYMOUS = "Anonymous" - USERNAME_AND_PASSWORD = "UsernameAndPassword" - WINDOWS_AZURE_HD_INSIGHT_SERVICE = "WindowsAzureHDInsightService" - -class PipelineReferenceType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Pipeline reference type. - """ - - PIPELINE_REFERENCE = "PipelineReference" - -class PluginCurrentState(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - - PREPARATION = "Preparation" - RESOURCE_ACQUISITION = "ResourceAcquisition" - QUEUED = "Queued" - SUBMISSION = "Submission" - MONITORING = "Monitoring" - CLEANUP = "Cleanup" - ENDED = "Ended" - -class PolybaseSettingsRejectType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Indicates whether the RejectValue property is specified as a literal value or a percentage. - """ - - VALUE = "value" - PERCENTAGE = "percentage" - -class PrestoAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The authentication mechanism used to connect to the Presto server. - """ - - ANONYMOUS = "Anonymous" - LDAP = "LDAP" - -class RecurrenceFrequency(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Enumerates possible frequency option for the schedule trigger. - """ - - NOT_SPECIFIED = "NotSpecified" - MINUTE = "Minute" - HOUR = "Hour" - DAY = "Day" - WEEK = "Week" - MONTH = "Month" - YEAR = "Year" - -class ResourceIdentityType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The type of managed identity for the workspace - """ - - NONE = "None" - SYSTEM_ASSIGNED = "SystemAssigned" - -class RestServiceAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Type of authentication used to connect to the REST service. - """ - - ANONYMOUS = "Anonymous" - BASIC = "Basic" - AAD_SERVICE_PRINCIPAL = "AadServicePrincipal" - MANAGED_SERVICE_IDENTITY = "ManagedServiceIdentity" - -class RunQueryFilterOperand(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Parameter name to be used for filter. The allowed operands to query pipeline runs are - PipelineName, RunStart, RunEnd and Status; to query activity runs are ActivityName, - ActivityRunStart, ActivityRunEnd, ActivityType and Status, and to query trigger runs are - TriggerName, TriggerRunTimestamp and Status. - """ - - PIPELINE_NAME = "PipelineName" - STATUS = "Status" - RUN_START = "RunStart" - RUN_END = "RunEnd" - ACTIVITY_NAME = "ActivityName" - ACTIVITY_RUN_START = "ActivityRunStart" - ACTIVITY_RUN_END = "ActivityRunEnd" - ACTIVITY_TYPE = "ActivityType" - TRIGGER_NAME = "TriggerName" - TRIGGER_RUN_TIMESTAMP = "TriggerRunTimestamp" - RUN_GROUP_ID = "RunGroupId" - LATEST_ONLY = "LatestOnly" - -class RunQueryFilterOperator(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Operator to be used for filter. - """ - - EQUALS = "Equals" - NOT_EQUALS = "NotEquals" - IN_ENUM = "In" - NOT_IN = "NotIn" - -class RunQueryOrder(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Sorting order of the parameter. - """ - - ASC = "ASC" - DESC = "DESC" - -class RunQueryOrderByField(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Parameter name to be used for order by. The allowed parameters to order by for pipeline runs - are PipelineName, RunStart, RunEnd and Status; for activity runs are ActivityName, - ActivityRunStart, ActivityRunEnd and Status; for trigger runs are TriggerName, - TriggerRunTimestamp and Status. - """ - - RUN_START = "RunStart" - RUN_END = "RunEnd" - PIPELINE_NAME = "PipelineName" - STATUS = "Status" - ACTIVITY_NAME = "ActivityName" - ACTIVITY_RUN_START = "ActivityRunStart" - ACTIVITY_RUN_END = "ActivityRunEnd" - TRIGGER_NAME = "TriggerName" - TRIGGER_RUN_TIMESTAMP = "TriggerRunTimestamp" - -class SalesforceSinkWriteBehavior(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The write behavior for the operation. Default is Insert. - """ - - INSERT = "Insert" - UPSERT = "Upsert" - -class SalesforceSourceReadBehavior(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The read behavior for the operation. Default is Query. - """ - - QUERY = "Query" - QUERY_ALL = "QueryAll" - -class SapCloudForCustomerSinkWriteBehavior(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The write behavior for the operation. Default is 'Insert'. - """ - - INSERT = "Insert" - UPDATE = "Update" - -class SapHanaAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The authentication type to be used to connect to the SAP HANA server. - """ - - BASIC = "Basic" - WINDOWS = "Windows" - -class SapHanaPartitionOption(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The partition mechanism that will be used for SAP HANA read in parallel. - """ - - NONE = "None" - PHYSICAL_PARTITIONS_OF_TABLE = "PhysicalPartitionsOfTable" - SAP_HANA_DYNAMIC_RANGE = "SapHanaDynamicRange" - -class SapTablePartitionOption(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The partition mechanism that will be used for SAP table read in parallel. - """ - - NONE = "None" - PARTITION_ON_INT = "PartitionOnInt" - PARTITION_ON_CALENDAR_YEAR = "PartitionOnCalendarYear" - PARTITION_ON_CALENDAR_MONTH = "PartitionOnCalendarMonth" - PARTITION_ON_CALENDAR_DATE = "PartitionOnCalendarDate" - PARTITION_ON_TIME = "PartitionOnTime" - -class SchedulerCurrentState(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - - QUEUED = "Queued" - SCHEDULED = "Scheduled" - ENDED = "Ended" - -class ServiceNowAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The authentication type to use. - """ - - BASIC = "Basic" - O_AUTH2 = "OAuth2" - -class SftpAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The authentication type to be used to connect to the FTP server. - """ - - BASIC = "Basic" - SSH_PUBLIC_KEY = "SshPublicKey" - -class SparkAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The authentication method used to access the Spark server. - """ - - ANONYMOUS = "Anonymous" - USERNAME = "Username" - USERNAME_AND_PASSWORD = "UsernameAndPassword" - WINDOWS_AZURE_HD_INSIGHT_SERVICE = "WindowsAzureHDInsightService" - -class SparkBatchJobResultType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The Spark batch job result. - """ - - UNCERTAIN = "Uncertain" - SUCCEEDED = "Succeeded" - FAILED = "Failed" - CANCELLED = "Cancelled" - -class SparkErrorSource(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - - SYSTEM = "System" - USER = "User" - UNKNOWN = "Unknown" - DEPENDENCY = "Dependency" - -class SparkJobReferenceType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Synapse spark job reference type. - """ - - SPARK_JOB_DEFINITION_REFERENCE = "SparkJobDefinitionReference" - -class SparkJobType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The job type. - """ - - SPARK_BATCH = "SparkBatch" - SPARK_SESSION = "SparkSession" - -class SparkServerType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The type of Spark server. - """ - - SHARK_SERVER = "SharkServer" - SHARK_SERVER2 = "SharkServer2" - SPARK_THRIFT_SERVER = "SparkThriftServer" - -class SparkThriftTransportProtocol(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The transport protocol to use in the Thrift layer. - """ - - BINARY = "Binary" - SASL = "SASL" - HTTP = "HTTP " - -class SqlConnectionType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The type of the connection. - """ - - SQL_ON_DEMAND = "SqlOnDemand" - SQL_POOL = "SqlPool" - -class SqlPartitionOption(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The partition mechanism that will be used for Sql read in parallel. - """ - - NONE = "None" - PHYSICAL_PARTITIONS_OF_TABLE = "PhysicalPartitionsOfTable" - DYNAMIC_RANGE = "DynamicRange" - -class SqlPoolReferenceType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """SQL pool reference type. - """ - - SQL_POOL_REFERENCE = "SqlPoolReference" - -class SqlScriptType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The type of the SQL script. - """ - - SQL_QUERY = "SqlQuery" - -class SsisLogLocationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The type of SSIS log location. - """ - - FILE = "File" - -class SsisPackageLocationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The type of SSIS package location. - """ - - SSISDB = "SSISDB" - FILE = "File" - INLINE_PACKAGE = "InlinePackage" - PACKAGE_STORE = "PackageStore" - -class StoredProcedureParameterType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Stored procedure parameter type. - """ - - STRING = "String" - INT = "Int" - INT64 = "Int64" - DECIMAL = "Decimal" - GUID = "Guid" - BOOLEAN = "Boolean" - DATE = "Date" - -class SybaseAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """AuthenticationType to be used for connection. - """ - - BASIC = "Basic" - WINDOWS = "Windows" - -class TeradataAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """AuthenticationType to be used for connection. - """ - - BASIC = "Basic" - WINDOWS = "Windows" - -class TeradataPartitionOption(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The partition mechanism that will be used for teradata read in parallel. - """ - - NONE = "None" - HASH = "Hash" - DYNAMIC_RANGE = "DynamicRange" - -class TriggerReferenceType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Trigger reference type. - """ - - TRIGGER_REFERENCE = "TriggerReference" - -class TriggerRunStatus(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Trigger run status. - """ - - SUCCEEDED = "Succeeded" - FAILED = "Failed" - INPROGRESS = "Inprogress" - -class TriggerRuntimeState(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Enumerates possible state of Triggers. - """ - - STARTED = "Started" - STOPPED = "Stopped" - DISABLED = "Disabled" - -class TumblingWindowFrequency(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Enumerates possible frequency option for the tumbling window trigger. - """ - - MINUTE = "Minute" - HOUR = "Hour" - MONTH = "Month" - -class Type(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Linked service reference type. - """ - - LINKED_SERVICE_REFERENCE = "LinkedServiceReference" - -class VariableType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Variable type. - """ - - STRING = "String" - BOOL = "Bool" - BOOLEAN = "Boolean" - ARRAY = "Array" - -class WebActivityMethod(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The list of HTTP methods supported by a WebActivity. - """ - - GET = "GET" - POST = "POST" - PUT = "PUT" - DELETE = "DELETE" - -class WebAuthenticationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """Type of authentication used to connect to the web table source. - """ - - BASIC = "Basic" - ANONYMOUS = "Anonymous" - CLIENT_CERTIFICATE = "ClientCertificate" - -class WebHookActivityMethod(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): - """The list of HTTP methods supported by a WebHook activity. - """ - - POST = "POST" diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_big_data_pools_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_big_data_pools_operations.py deleted file mode 100644 index bc427831f4b3..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_big_data_pools_operations.py +++ /dev/null @@ -1,152 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Optional, TypeVar - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class BigDataPoolsOperations(object): - """BigDataPoolsOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def list( - self, - **kwargs # type: Any - ): - # type: (...) -> "_models.BigDataPoolResourceInfoListResult" - """List Big Data Pools. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: BigDataPoolResourceInfoListResult, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.BigDataPoolResourceInfoListResult - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.BigDataPoolResourceInfoListResult"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.list.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('BigDataPoolResourceInfoListResult', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - list.metadata = {'url': '/bigDataPools'} # type: ignore - - def get( - self, - big_data_pool_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> "_models.BigDataPoolResourceInfo" - """Get Big Data Pool. - - :param big_data_pool_name: The Big Data Pool name. - :type big_data_pool_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: BigDataPoolResourceInfo, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.BigDataPoolResourceInfo - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.BigDataPoolResourceInfo"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'bigDataPoolName': self._serialize.url("big_data_pool_name", big_data_pool_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('BigDataPoolResourceInfo', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get.metadata = {'url': '/bigDataPools/{bigDataPoolName}'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_data_flow_debug_session_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_data_flow_debug_session_operations.py deleted file mode 100644 index 8c443f2d704e..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_data_flow_debug_session_operations.py +++ /dev/null @@ -1,474 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.paging import ItemPaged -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse -from azure.core.polling import LROPoller, NoPolling, PollingMethod -from azure.core.polling.base_polling import LROBasePolling - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Iterable, Optional, TypeVar, Union - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class DataFlowDebugSessionOperations(object): - """DataFlowDebugSessionOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def _create_data_flow_debug_session_initial( - self, - request, # type: "_models.CreateDataFlowDebugSessionRequest" - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.CreateDataFlowDebugSessionResponse"] - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.CreateDataFlowDebugSessionResponse"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._create_data_flow_debug_session_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(request, 'CreateDataFlowDebugSessionRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - response_headers = {} - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('CreateDataFlowDebugSessionResponse', pipeline_response) - - if response.status_code == 202: - response_headers['location']=self._deserialize('str', response.headers.get('location')) - - if cls: - return cls(pipeline_response, deserialized, response_headers) - - return deserialized - _create_data_flow_debug_session_initial.metadata = {'url': '/createDataFlowDebugSession'} # type: ignore - - def begin_create_data_flow_debug_session( - self, - request, # type: "_models.CreateDataFlowDebugSessionRequest" - **kwargs # type: Any - ): - # type: (...) -> LROPoller["_models.CreateDataFlowDebugSessionResponse"] - """Creates a data flow debug session. - - :param request: Data flow debug session definition. - :type request: ~azure.synapse.artifacts.v2021_06_01_preview.models.CreateDataFlowDebugSessionRequest - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either CreateDataFlowDebugSessionResponse or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.CreateDataFlowDebugSessionResponse] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.CreateDataFlowDebugSessionResponse"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._create_data_flow_debug_session_initial( - request=request, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('CreateDataFlowDebugSessionResponse', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create_data_flow_debug_session.metadata = {'url': '/createDataFlowDebugSession'} # type: ignore - - def query_data_flow_debug_sessions_by_workspace( - self, - **kwargs # type: Any - ): - # type: (...) -> Iterable["_models.QueryDataFlowDebugSessionsResponse"] - """Query all active data flow debug sessions. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either QueryDataFlowDebugSessionsResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.QueryDataFlowDebugSessionsResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.QueryDataFlowDebugSessionsResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.query_data_flow_debug_sessions_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - def extract_data(pipeline_response): - deserialized = self._deserialize('QueryDataFlowDebugSessionsResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, iter(list_of_elem) - - def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return ItemPaged( - get_next, extract_data - ) - query_data_flow_debug_sessions_by_workspace.metadata = {'url': '/queryDataFlowDebugSessions'} # type: ignore - - def add_data_flow( - self, - request, # type: "_models.DataFlowDebugPackage" - **kwargs # type: Any - ): - # type: (...) -> "_models.AddDataFlowToDebugSessionResponse" - """Add a data flow into debug session. - - :param request: Data flow debug session definition with debug content. - :type request: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowDebugPackage - :keyword callable cls: A custom type or function that will be passed the direct response - :return: AddDataFlowToDebugSessionResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.AddDataFlowToDebugSessionResponse - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.AddDataFlowToDebugSessionResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self.add_data_flow.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(request, 'DataFlowDebugPackage') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('AddDataFlowToDebugSessionResponse', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - add_data_flow.metadata = {'url': '/addDataFlowToDebugSession'} # type: ignore - - def delete_data_flow_debug_session( - self, - request, # type: "_models.DeleteDataFlowDebugSessionRequest" - **kwargs # type: Any - ): - # type: (...) -> None - """Deletes a data flow debug session. - - :param request: Data flow debug session definition for deletion. - :type request: ~azure.synapse.artifacts.v2021_06_01_preview.models.DeleteDataFlowDebugSessionRequest - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self.delete_data_flow_debug_session.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(request, 'DeleteDataFlowDebugSessionRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - delete_data_flow_debug_session.metadata = {'url': '/deleteDataFlowDebugSession'} # type: ignore - - def _execute_command_initial( - self, - request, # type: "_models.DataFlowDebugCommandRequest" - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.DataFlowDebugCommandResponse"] - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.DataFlowDebugCommandResponse"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._execute_command_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(request, 'DataFlowDebugCommandRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - response_headers = {} - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('DataFlowDebugCommandResponse', pipeline_response) - - if response.status_code == 202: - response_headers['location']=self._deserialize('str', response.headers.get('location')) - - if cls: - return cls(pipeline_response, deserialized, response_headers) - - return deserialized - _execute_command_initial.metadata = {'url': '/executeDataFlowDebugCommand'} # type: ignore - - def begin_execute_command( - self, - request, # type: "_models.DataFlowDebugCommandRequest" - **kwargs # type: Any - ): - # type: (...) -> LROPoller["_models.DataFlowDebugCommandResponse"] - """Execute a data flow debug command. - - :param request: Data flow debug command definition. - :type request: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowDebugCommandRequest - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either DataFlowDebugCommandResponse or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowDebugCommandResponse] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.DataFlowDebugCommandResponse"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._execute_command_initial( - request=request, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('DataFlowDebugCommandResponse', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_execute_command.metadata = {'url': '/executeDataFlowDebugCommand'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_data_flow_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_data_flow_operations.py deleted file mode 100644 index 58a963cce999..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_data_flow_operations.py +++ /dev/null @@ -1,533 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.paging import ItemPaged -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse -from azure.core.polling import LROPoller, NoPolling, PollingMethod -from azure.core.polling.base_polling import LROBasePolling - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Iterable, Optional, TypeVar, Union - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class DataFlowOperations(object): - """DataFlowOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def _create_or_update_data_flow_initial( - self, - data_flow_name, # type: str - properties, # type: "_models.DataFlow" - if_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.DataFlowResource"] - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.DataFlowResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _data_flow = _models.DataFlowResource(properties=properties) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._create_or_update_data_flow_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'dataFlowName': self._serialize.url("data_flow_name", data_flow_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_match is not None: - header_parameters['If-Match'] = self._serialize.header("if_match", if_match, 'str') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_data_flow, 'DataFlowResource') - body_content_kwargs['content'] = body_content - request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('DataFlowResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _create_or_update_data_flow_initial.metadata = {'url': '/dataflows/{dataFlowName}'} # type: ignore - - def begin_create_or_update_data_flow( - self, - data_flow_name, # type: str - properties, # type: "_models.DataFlow" - if_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> LROPoller["_models.DataFlowResource"] - """Creates or updates a data flow. - - :param data_flow_name: The data flow name. - :type data_flow_name: str - :param properties: Data flow properties. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlow - :param if_match: ETag of the data flow entity. Should only be specified for update, for which - it should match existing entity or can be * for unconditional update. - :type if_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either DataFlowResource or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowResource] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.DataFlowResource"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._create_or_update_data_flow_initial( - data_flow_name=data_flow_name, - properties=properties, - if_match=if_match, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('DataFlowResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'dataFlowName': self._serialize.url("data_flow_name", data_flow_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create_or_update_data_flow.metadata = {'url': '/dataflows/{dataFlowName}'} # type: ignore - - def get_data_flow( - self, - data_flow_name, # type: str - if_none_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> "_models.DataFlowResource" - """Gets a data flow. - - :param data_flow_name: The data flow name. - :type data_flow_name: str - :param if_none_match: ETag of the data flow entity. Should only be specified for get. If the - ETag matches the existing entity tag, or if * was provided, then no content will be returned. - :type if_none_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: DataFlowResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowResource - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.DataFlowResource"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_data_flow.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'dataFlowName': self._serialize.url("data_flow_name", data_flow_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_none_match is not None: - header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('DataFlowResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_data_flow.metadata = {'url': '/dataflows/{dataFlowName}'} # type: ignore - - def _delete_data_flow_initial( - self, - data_flow_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._delete_data_flow_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'dataFlowName': self._serialize.url("data_flow_name", data_flow_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.delete(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _delete_data_flow_initial.metadata = {'url': '/dataflows/{dataFlowName}'} # type: ignore - - def begin_delete_data_flow( - self, - data_flow_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> LROPoller[None] - """Deletes a data flow. - - :param data_flow_name: The data flow name. - :type data_flow_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._delete_data_flow_initial( - data_flow_name=data_flow_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'dataFlowName': self._serialize.url("data_flow_name", data_flow_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_delete_data_flow.metadata = {'url': '/dataflows/{dataFlowName}'} # type: ignore - - def _rename_data_flow_initial( - self, - data_flow_name, # type: str - new_name=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> None - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _request = _models.ArtifactRenameRequest(new_name=new_name) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._rename_data_flow_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'dataFlowName': self._serialize.url("data_flow_name", data_flow_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_request, 'ArtifactRenameRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _rename_data_flow_initial.metadata = {'url': '/dataflows/{dataFlowName}/rename'} # type: ignore - - def begin_rename_data_flow( - self, - data_flow_name, # type: str - new_name=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> LROPoller[None] - """Renames a dataflow. - - :param data_flow_name: The data flow name. - :type data_flow_name: str - :param new_name: New name of the artifact. - :type new_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._rename_data_flow_initial( - data_flow_name=data_flow_name, - new_name=new_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'dataFlowName': self._serialize.url("data_flow_name", data_flow_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_rename_data_flow.metadata = {'url': '/dataflows/{dataFlowName}/rename'} # type: ignore - - def get_data_flows_by_workspace( - self, - **kwargs # type: Any - ): - # type: (...) -> Iterable["_models.DataFlowListResponse"] - """Lists data flows. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either DataFlowListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.DataFlowListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.DataFlowListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.get_data_flows_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - def extract_data(pipeline_response): - deserialized = self._deserialize('DataFlowListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, iter(list_of_elem) - - def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return ItemPaged( - get_next, extract_data - ) - get_data_flows_by_workspace.metadata = {'url': '/dataflows'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_dataset_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_dataset_operations.py deleted file mode 100644 index 8cc21d8c7b69..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_dataset_operations.py +++ /dev/null @@ -1,535 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.paging import ItemPaged -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse -from azure.core.polling import LROPoller, NoPolling, PollingMethod -from azure.core.polling.base_polling import LROBasePolling - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Iterable, Optional, TypeVar, Union - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class DatasetOperations(object): - """DatasetOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get_datasets_by_workspace( - self, - **kwargs # type: Any - ): - # type: (...) -> Iterable["_models.DatasetListResponse"] - """Lists datasets. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either DatasetListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.DatasetListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.get_datasets_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - def extract_data(pipeline_response): - deserialized = self._deserialize('DatasetListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, iter(list_of_elem) - - def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return ItemPaged( - get_next, extract_data - ) - get_datasets_by_workspace.metadata = {'url': '/datasets'} # type: ignore - - def _create_or_update_dataset_initial( - self, - dataset_name, # type: str - properties, # type: "_models.Dataset" - if_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.DatasetResource"] - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.DatasetResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _dataset = _models.DatasetResource(properties=properties) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._create_or_update_dataset_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'datasetName': self._serialize.url("dataset_name", dataset_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_match is not None: - header_parameters['If-Match'] = self._serialize.header("if_match", if_match, 'str') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_dataset, 'DatasetResource') - body_content_kwargs['content'] = body_content - request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('DatasetResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _create_or_update_dataset_initial.metadata = {'url': '/datasets/{datasetName}'} # type: ignore - - def begin_create_or_update_dataset( - self, - dataset_name, # type: str - properties, # type: "_models.Dataset" - if_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> LROPoller["_models.DatasetResource"] - """Creates or updates a dataset. - - :param dataset_name: The dataset name. - :type dataset_name: str - :param properties: Dataset properties. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.Dataset - :param if_match: ETag of the dataset entity. Should only be specified for update, for which it - should match existing entity or can be * for unconditional update. - :type if_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either DatasetResource or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetResource] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.DatasetResource"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._create_or_update_dataset_initial( - dataset_name=dataset_name, - properties=properties, - if_match=if_match, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('DatasetResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'datasetName': self._serialize.url("dataset_name", dataset_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create_or_update_dataset.metadata = {'url': '/datasets/{datasetName}'} # type: ignore - - def get_dataset( - self, - dataset_name, # type: str - if_none_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.DatasetResource"] - """Gets a dataset. - - :param dataset_name: The dataset name. - :type dataset_name: str - :param if_none_match: ETag of the dataset entity. Should only be specified for get. If the ETag - matches the existing entity tag, or if * was provided, then no content will be returned. - :type if_none_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: DatasetResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.DatasetResource or None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.DatasetResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_dataset.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'datasetName': self._serialize.url("dataset_name", dataset_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_none_match is not None: - header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 304]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('DatasetResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_dataset.metadata = {'url': '/datasets/{datasetName}'} # type: ignore - - def _delete_dataset_initial( - self, - dataset_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._delete_dataset_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'datasetName': self._serialize.url("dataset_name", dataset_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.delete(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _delete_dataset_initial.metadata = {'url': '/datasets/{datasetName}'} # type: ignore - - def begin_delete_dataset( - self, - dataset_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> LROPoller[None] - """Deletes a dataset. - - :param dataset_name: The dataset name. - :type dataset_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._delete_dataset_initial( - dataset_name=dataset_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'datasetName': self._serialize.url("dataset_name", dataset_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_delete_dataset.metadata = {'url': '/datasets/{datasetName}'} # type: ignore - - def _rename_dataset_initial( - self, - dataset_name, # type: str - new_name=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> None - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _request = _models.ArtifactRenameRequest(new_name=new_name) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._rename_dataset_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'datasetName': self._serialize.url("dataset_name", dataset_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_request, 'ArtifactRenameRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _rename_dataset_initial.metadata = {'url': '/datasets/{datasetName}/rename'} # type: ignore - - def begin_rename_dataset( - self, - dataset_name, # type: str - new_name=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> LROPoller[None] - """Renames a dataset. - - :param dataset_name: The dataset name. - :type dataset_name: str - :param new_name: New name of the artifact. - :type new_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._rename_dataset_initial( - dataset_name=dataset_name, - new_name=new_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'datasetName': self._serialize.url("dataset_name", dataset_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_rename_dataset.metadata = {'url': '/datasets/{datasetName}/rename'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_integration_runtimes_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_integration_runtimes_operations.py deleted file mode 100644 index 0fd7307f625e..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_integration_runtimes_operations.py +++ /dev/null @@ -1,152 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Optional, TypeVar - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class IntegrationRuntimesOperations(object): - """IntegrationRuntimesOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def list( - self, - **kwargs # type: Any - ): - # type: (...) -> "_models.IntegrationRuntimeListResponse" - """List Integration Runtimes. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: IntegrationRuntimeListResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeListResponse - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.IntegrationRuntimeListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.list.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('IntegrationRuntimeListResponse', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - list.metadata = {'url': '/integrationRuntimes'} # type: ignore - - def get( - self, - integration_runtime_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> "_models.IntegrationRuntimeResource" - """Get Integration Runtime. - - :param integration_runtime_name: The Integration Runtime name. - :type integration_runtime_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: IntegrationRuntimeResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.IntegrationRuntimeResource - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.IntegrationRuntimeResource"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'integrationRuntimeName': self._serialize.url("integration_runtime_name", integration_runtime_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('IntegrationRuntimeResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get.metadata = {'url': '/integrationRuntimes/{integrationRuntimeName}'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_library_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_library_operations.py deleted file mode 100644 index d79acf6c8213..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_library_operations.py +++ /dev/null @@ -1,649 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.paging import ItemPaged -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse -from azure.core.polling import LROPoller, NoPolling, PollingMethod -from azure.core.polling.base_polling import LROBasePolling - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, IO, Iterable, Optional, TypeVar, Union - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class LibraryOperations(object): - """LibraryOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def list( - self, - **kwargs # type: Any - ): - # type: (...) -> Iterable["_models.LibraryListResponse"] - """Lists Library. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either LibraryListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.LibraryListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.LibraryListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.list.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - def extract_data(pipeline_response): - deserialized = self._deserialize('LibraryListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, iter(list_of_elem) - - def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return ItemPaged( - get_next, extract_data - ) - list.metadata = {'url': '/libraries'} # type: ignore - - def _flush_initial( - self, - library_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.LibraryResourceInfo"] - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.LibraryResourceInfo"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._flush_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'libraryName': self._serialize.url("library_name", library_name, 'str', max_length=100, min_length=0), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 202: - deserialized = self._deserialize('LibraryResourceInfo', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _flush_initial.metadata = {'url': '/libraries/{libraryName}/flush'} # type: ignore - - def begin_flush( - self, - library_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> LROPoller["_models.LibraryResourceInfo"] - """Flush Library. - - :param library_name: file name to upload. Minimum length of the filename should be 1 excluding - the extension length. - :type library_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.LibraryResourceInfo"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._flush_initial( - library_name=library_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('LibraryResourceInfo', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'libraryName': self._serialize.url("library_name", library_name, 'str', max_length=100, min_length=0), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_flush.metadata = {'url': '/libraries/{libraryName}/flush'} # type: ignore - - def get_operation_result( - self, - operation_id, # type: str - **kwargs # type: Any - ): - # type: (...) -> Union["_models.LibraryResource", "_models.OperationResult"] - """Get Operation result for Library. - - :param operation_id: operation id for which status is requested. - :type operation_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: LibraryResource or OperationResult, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.LibraryResource or ~azure.synapse.artifacts.v2021_06_01_preview.models.OperationResult - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[Union["_models.LibraryResource", "_models.OperationResult"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_operation_result.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'operationId': self._serialize.url("operation_id", operation_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if response.status_code == 200: - deserialized = self._deserialize('LibraryResource', pipeline_response) - - if response.status_code == 202: - deserialized = self._deserialize('OperationResult', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_operation_result.metadata = {'url': '/libraryOperationResults/{operationId}'} # type: ignore - - def _delete_initial( - self, - library_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.LibraryResourceInfo"] - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.LibraryResourceInfo"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._delete_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'libraryName': self._serialize.url("library_name", library_name, 'str', max_length=100, min_length=0), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.delete(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202, 409]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 202: - deserialized = self._deserialize('LibraryResourceInfo', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _delete_initial.metadata = {'url': '/libraries/{libraryName}'} # type: ignore - - def begin_delete( - self, - library_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> LROPoller["_models.LibraryResourceInfo"] - """Delete Library. - - :param library_name: file name to upload. Minimum length of the filename should be 1 excluding - the extension length. - :type library_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.LibraryResourceInfo"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._delete_initial( - library_name=library_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('LibraryResourceInfo', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'libraryName': self._serialize.url("library_name", library_name, 'str', max_length=100, min_length=0), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_delete.metadata = {'url': '/libraries/{libraryName}'} # type: ignore - - def get( - self, - library_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.LibraryResource"] - """Get Library. - - :param library_name: file name to upload. Minimum length of the filename should be 1 excluding - the extension length. - :type library_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: LibraryResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.LibraryResource or None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.LibraryResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'libraryName': self._serialize.url("library_name", library_name, 'str', max_length=100, min_length=0), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 304]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('LibraryResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get.metadata = {'url': '/libraries/{libraryName}'} # type: ignore - - def _create_initial( - self, - library_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.LibraryResourceInfo"] - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.LibraryResourceInfo"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._create_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'libraryName': self._serialize.url("library_name", library_name, 'str', max_length=100, min_length=0), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.put(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 202: - deserialized = self._deserialize('LibraryResourceInfo', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _create_initial.metadata = {'url': '/libraries/{libraryName}'} # type: ignore - - def begin_create( - self, - library_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> LROPoller["_models.LibraryResourceInfo"] - """Creates a library with the library name. - - :param library_name: file name to upload. Minimum length of the filename should be 1 excluding - the extension length. - :type library_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.LibraryResourceInfo"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._create_initial( - library_name=library_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('LibraryResourceInfo', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'libraryName': self._serialize.url("library_name", library_name, 'str', max_length=100, min_length=0), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create.metadata = {'url': '/libraries/{libraryName}'} # type: ignore - - def append( - self, - library_name, # type: str - content, # type: IO - blob_condition_append_position=None, # type: Optional[int] - **kwargs # type: Any - ): - # type: (...) -> None - """Append the content to the library resource created using the create operation. The maximum - content size is 4MiB. Content larger than 4MiB must be appended in 4MiB chunks. - - :param library_name: file name to upload. Minimum length of the filename should be 1 excluding - the extension length. - :type library_name: str - :param content: Library file chunk. - :type content: IO - :param blob_condition_append_position: Set this header to a byte offset at which the block is - expected to be appended. The request succeeds only if the current offset matches this value. - Otherwise, the request fails with the AppendPositionConditionNotMet error (HTTP status code 412 - – Precondition Failed). - :type blob_condition_append_position: long - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - comp = "appendblock" - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/octet-stream") - accept = "application/json" - - # Construct URL - url = self.append.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'libraryName': self._serialize.url("library_name", library_name, 'str', max_length=100, min_length=0), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['comp'] = self._serialize.query("comp", comp, 'str') - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if blob_condition_append_position is not None: - header_parameters['x-ms-blob-condition-appendpos'] = self._serialize.header("blob_condition_append_position", blob_condition_append_position, 'long') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content_kwargs['stream_content'] = content - request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [201]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - append.metadata = {'url': '/libraries/{libraryName}'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_linked_service_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_linked_service_operations.py deleted file mode 100644 index f8f6507ff15f..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_linked_service_operations.py +++ /dev/null @@ -1,536 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.paging import ItemPaged -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse -from azure.core.polling import LROPoller, NoPolling, PollingMethod -from azure.core.polling.base_polling import LROBasePolling - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Iterable, Optional, TypeVar, Union - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class LinkedServiceOperations(object): - """LinkedServiceOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get_linked_services_by_workspace( - self, - **kwargs # type: Any - ): - # type: (...) -> Iterable["_models.LinkedServiceListResponse"] - """Lists linked services. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either LinkedServiceListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.LinkedServiceListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.get_linked_services_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - def extract_data(pipeline_response): - deserialized = self._deserialize('LinkedServiceListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, iter(list_of_elem) - - def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return ItemPaged( - get_next, extract_data - ) - get_linked_services_by_workspace.metadata = {'url': '/linkedservices'} # type: ignore - - def _create_or_update_linked_service_initial( - self, - linked_service_name, # type: str - properties, # type: "_models.LinkedService" - if_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.LinkedServiceResource"] - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.LinkedServiceResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _linked_service = _models.LinkedServiceResource(properties=properties) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._create_or_update_linked_service_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'linkedServiceName': self._serialize.url("linked_service_name", linked_service_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_match is not None: - header_parameters['If-Match'] = self._serialize.header("if_match", if_match, 'str') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_linked_service, 'LinkedServiceResource') - body_content_kwargs['content'] = body_content - request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('LinkedServiceResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _create_or_update_linked_service_initial.metadata = {'url': '/linkedservices/{linkedServiceName}'} # type: ignore - - def begin_create_or_update_linked_service( - self, - linked_service_name, # type: str - properties, # type: "_models.LinkedService" - if_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> LROPoller["_models.LinkedServiceResource"] - """Creates or updates a linked service. - - :param linked_service_name: The linked service name. - :type linked_service_name: str - :param properties: Properties of linked service. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedService - :param if_match: ETag of the linkedService entity. Should only be specified for update, for - which it should match existing entity or can be * for unconditional update. - :type if_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either LinkedServiceResource or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceResource] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.LinkedServiceResource"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._create_or_update_linked_service_initial( - linked_service_name=linked_service_name, - properties=properties, - if_match=if_match, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('LinkedServiceResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'linkedServiceName': self._serialize.url("linked_service_name", linked_service_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create_or_update_linked_service.metadata = {'url': '/linkedservices/{linkedServiceName}'} # type: ignore - - def get_linked_service( - self, - linked_service_name, # type: str - if_none_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.LinkedServiceResource"] - """Gets a linked service. - - :param linked_service_name: The linked service name. - :type linked_service_name: str - :param if_none_match: ETag of the linked service entity. Should only be specified for get. If - the ETag matches the existing entity tag, or if * was provided, then no content will be - returned. - :type if_none_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: LinkedServiceResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.LinkedServiceResource or None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.LinkedServiceResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_linked_service.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'linkedServiceName': self._serialize.url("linked_service_name", linked_service_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_none_match is not None: - header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 304]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('LinkedServiceResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_linked_service.metadata = {'url': '/linkedservices/{linkedServiceName}'} # type: ignore - - def _delete_linked_service_initial( - self, - linked_service_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._delete_linked_service_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'linkedServiceName': self._serialize.url("linked_service_name", linked_service_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.delete(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _delete_linked_service_initial.metadata = {'url': '/linkedservices/{linkedServiceName}'} # type: ignore - - def begin_delete_linked_service( - self, - linked_service_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> LROPoller[None] - """Deletes a linked service. - - :param linked_service_name: The linked service name. - :type linked_service_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._delete_linked_service_initial( - linked_service_name=linked_service_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'linkedServiceName': self._serialize.url("linked_service_name", linked_service_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_delete_linked_service.metadata = {'url': '/linkedservices/{linkedServiceName}'} # type: ignore - - def _rename_linked_service_initial( - self, - linked_service_name, # type: str - new_name=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> None - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _request = _models.ArtifactRenameRequest(new_name=new_name) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._rename_linked_service_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'linkedServiceName': self._serialize.url("linked_service_name", linked_service_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_request, 'ArtifactRenameRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _rename_linked_service_initial.metadata = {'url': '/linkedservices/{linkedServiceName}/rename'} # type: ignore - - def begin_rename_linked_service( - self, - linked_service_name, # type: str - new_name=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> LROPoller[None] - """Renames a linked service. - - :param linked_service_name: The linked service name. - :type linked_service_name: str - :param new_name: New name of the artifact. - :type new_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._rename_linked_service_initial( - linked_service_name=linked_service_name, - new_name=new_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'linkedServiceName': self._serialize.url("linked_service_name", linked_service_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_rename_linked_service.metadata = {'url': '/linkedservices/{linkedServiceName}/rename'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_notebook_operation_result_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_notebook_operation_result_operations.py deleted file mode 100644 index cae1f4c2661e..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_notebook_operation_result_operations.py +++ /dev/null @@ -1,97 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Optional, TypeVar - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class NotebookOperationResultOperations(object): - """NotebookOperationResultOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get( - self, - operation_id, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - """Get notebook operation result. - - :param operation_id: Operation ID. - :type operation_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'operationId': self._serialize.url("operation_id", operation_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 201, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - get.metadata = {'url': '/notebookOperationResults/{operationId}'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_notebook_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_notebook_operations.py deleted file mode 100644 index 528ea15a5de5..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_notebook_operations.py +++ /dev/null @@ -1,605 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.paging import ItemPaged -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse -from azure.core.polling import LROPoller, NoPolling, PollingMethod -from azure.core.polling.base_polling import LROBasePolling - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Iterable, Optional, TypeVar, Union - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class NotebookOperations(object): - """NotebookOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get_notebooks_by_workspace( - self, - **kwargs # type: Any - ): - # type: (...) -> Iterable["_models.NotebookListResponse"] - """Lists Notebooks. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either NotebookListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.NotebookListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.get_notebooks_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - def extract_data(pipeline_response): - deserialized = self._deserialize('NotebookListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, iter(list_of_elem) - - def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return ItemPaged( - get_next, extract_data - ) - get_notebooks_by_workspace.metadata = {'url': '/notebooks'} # type: ignore - - def get_notebook_summary_by_work_space( - self, - **kwargs # type: Any - ): - # type: (...) -> Iterable["_models.NotebookListResponse"] - """Lists a summary of Notebooks. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either NotebookListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.NotebookListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.get_notebook_summary_by_work_space.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - def extract_data(pipeline_response): - deserialized = self._deserialize('NotebookListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, iter(list_of_elem) - - def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return ItemPaged( - get_next, extract_data - ) - get_notebook_summary_by_work_space.metadata = {'url': '/notebooksSummary'} # type: ignore - - def _create_or_update_notebook_initial( - self, - notebook_name, # type: str - notebook, # type: "_models.NotebookResource" - if_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.NotebookResource"] - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.NotebookResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._create_or_update_notebook_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'notebookName': self._serialize.url("notebook_name", notebook_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_match is not None: - header_parameters['If-Match'] = self._serialize.header("if_match", if_match, 'str') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(notebook, 'NotebookResource') - body_content_kwargs['content'] = body_content - request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('NotebookResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _create_or_update_notebook_initial.metadata = {'url': '/notebooks/{notebookName}'} # type: ignore - - def begin_create_or_update_notebook( - self, - notebook_name, # type: str - notebook, # type: "_models.NotebookResource" - if_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> LROPoller["_models.NotebookResource"] - """Creates or updates a Note Book. - - :param notebook_name: The notebook name. - :type notebook_name: str - :param notebook: Note book resource definition. - :type notebook: ~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookResource - :param if_match: ETag of the Note book entity. Should only be specified for update, for which - it should match existing entity or can be * for unconditional update. - :type if_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either NotebookResource or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookResource] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.NotebookResource"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._create_or_update_notebook_initial( - notebook_name=notebook_name, - notebook=notebook, - if_match=if_match, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('NotebookResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'notebookName': self._serialize.url("notebook_name", notebook_name, 'str'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create_or_update_notebook.metadata = {'url': '/notebooks/{notebookName}'} # type: ignore - - def get_notebook( - self, - notebook_name, # type: str - if_none_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.NotebookResource"] - """Gets a Note Book. - - :param notebook_name: The notebook name. - :type notebook_name: str - :param if_none_match: ETag of the Notebook entity. Should only be specified for get. If the - ETag matches the existing entity tag, or if * was provided, then no content will be returned. - :type if_none_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: NotebookResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.NotebookResource or None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.NotebookResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_notebook.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'notebookName': self._serialize.url("notebook_name", notebook_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_none_match is not None: - header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 304]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('NotebookResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_notebook.metadata = {'url': '/notebooks/{notebookName}'} # type: ignore - - def _delete_notebook_initial( - self, - notebook_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._delete_notebook_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'notebookName': self._serialize.url("notebook_name", notebook_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.delete(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _delete_notebook_initial.metadata = {'url': '/notebooks/{notebookName}'} # type: ignore - - def begin_delete_notebook( - self, - notebook_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> LROPoller[None] - """Deletes a Note book. - - :param notebook_name: The notebook name. - :type notebook_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._delete_notebook_initial( - notebook_name=notebook_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'notebookName': self._serialize.url("notebook_name", notebook_name, 'str'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_delete_notebook.metadata = {'url': '/notebooks/{notebookName}'} # type: ignore - - def _rename_notebook_initial( - self, - notebook_name, # type: str - new_name=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> None - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _request = _models.ArtifactRenameRequest(new_name=new_name) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._rename_notebook_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'notebookName': self._serialize.url("notebook_name", notebook_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_request, 'ArtifactRenameRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _rename_notebook_initial.metadata = {'url': '/notebooks/{notebookName}/rename'} # type: ignore - - def begin_rename_notebook( - self, - notebook_name, # type: str - new_name=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> LROPoller[None] - """Renames a notebook. - - :param notebook_name: The notebook name. - :type notebook_name: str - :param new_name: New name of the artifact. - :type new_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._rename_notebook_initial( - notebook_name=notebook_name, - new_name=new_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'notebookName': self._serialize.url("notebook_name", notebook_name, 'str'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_rename_notebook.metadata = {'url': '/notebooks/{notebookName}/rename'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_operation_result_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_operation_result_operations.py deleted file mode 100644 index 72ee864a43b3..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_operation_result_operations.py +++ /dev/null @@ -1,97 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Optional, TypeVar - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class OperationResultOperations(object): - """OperationResultOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get( - self, - operation_id, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - """Get operation result. - - :param operation_id: Operation ID. - :type operation_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'operationId': self._serialize.url("operation_id", operation_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 201, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - get.metadata = {'url': '/operationResults/{operationId}'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_operation_status_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_operation_status_operations.py deleted file mode 100644 index 32219fdcc654..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_operation_status_operations.py +++ /dev/null @@ -1,97 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Optional, TypeVar - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class OperationStatusOperations(object): - """OperationStatusOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get( - self, - operation_id, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - """Get operation status. - - :param operation_id: Operation ID. - :type operation_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'operationId': self._serialize.url("operation_id", operation_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 201, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - get.metadata = {'url': '/operationStatuses/{operationId}'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_pipeline_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_pipeline_operations.py deleted file mode 100644 index cec9861bc0b3..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_pipeline_operations.py +++ /dev/null @@ -1,619 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.paging import ItemPaged -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse -from azure.core.polling import LROPoller, NoPolling, PollingMethod -from azure.core.polling.base_polling import LROBasePolling - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Iterable, Optional, TypeVar, Union - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class PipelineOperations(object): - """PipelineOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get_pipelines_by_workspace( - self, - **kwargs # type: Any - ): - # type: (...) -> Iterable["_models.PipelineListResponse"] - """Lists pipelines. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either PipelineListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.PipelineListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.get_pipelines_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - def extract_data(pipeline_response): - deserialized = self._deserialize('PipelineListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, iter(list_of_elem) - - def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return ItemPaged( - get_next, extract_data - ) - get_pipelines_by_workspace.metadata = {'url': '/pipelines'} # type: ignore - - def _create_or_update_pipeline_initial( - self, - pipeline_name, # type: str - pipeline, # type: "_models.PipelineResource" - if_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.PipelineResource"] - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.PipelineResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._create_or_update_pipeline_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'pipelineName': self._serialize.url("pipeline_name", pipeline_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_match is not None: - header_parameters['If-Match'] = self._serialize.header("if_match", if_match, 'str') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(pipeline, 'PipelineResource') - body_content_kwargs['content'] = body_content - request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('PipelineResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _create_or_update_pipeline_initial.metadata = {'url': '/pipelines/{pipelineName}'} # type: ignore - - def begin_create_or_update_pipeline( - self, - pipeline_name, # type: str - pipeline, # type: "_models.PipelineResource" - if_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> LROPoller["_models.PipelineResource"] - """Creates or updates a pipeline. - - :param pipeline_name: The pipeline name. - :type pipeline_name: str - :param pipeline: Pipeline resource definition. - :type pipeline: ~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineResource - :param if_match: ETag of the pipeline entity. Should only be specified for update, for which - it should match existing entity or can be * for unconditional update. - :type if_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either PipelineResource or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineResource] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.PipelineResource"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._create_or_update_pipeline_initial( - pipeline_name=pipeline_name, - pipeline=pipeline, - if_match=if_match, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('PipelineResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'pipelineName': self._serialize.url("pipeline_name", pipeline_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create_or_update_pipeline.metadata = {'url': '/pipelines/{pipelineName}'} # type: ignore - - def get_pipeline( - self, - pipeline_name, # type: str - if_none_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.PipelineResource"] - """Gets a pipeline. - - :param pipeline_name: The pipeline name. - :type pipeline_name: str - :param if_none_match: ETag of the pipeline entity. Should only be specified for get. If the - ETag matches the existing entity tag, or if * was provided, then no content will be returned. - :type if_none_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: PipelineResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineResource or None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.PipelineResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_pipeline.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'pipelineName': self._serialize.url("pipeline_name", pipeline_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_none_match is not None: - header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 304]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('PipelineResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_pipeline.metadata = {'url': '/pipelines/{pipelineName}'} # type: ignore - - def _delete_pipeline_initial( - self, - pipeline_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._delete_pipeline_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'pipelineName': self._serialize.url("pipeline_name", pipeline_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.delete(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _delete_pipeline_initial.metadata = {'url': '/pipelines/{pipelineName}'} # type: ignore - - def begin_delete_pipeline( - self, - pipeline_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> LROPoller[None] - """Deletes a pipeline. - - :param pipeline_name: The pipeline name. - :type pipeline_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._delete_pipeline_initial( - pipeline_name=pipeline_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'pipelineName': self._serialize.url("pipeline_name", pipeline_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_delete_pipeline.metadata = {'url': '/pipelines/{pipelineName}'} # type: ignore - - def _rename_pipeline_initial( - self, - pipeline_name, # type: str - new_name=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> None - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _request = _models.ArtifactRenameRequest(new_name=new_name) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._rename_pipeline_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'pipelineName': self._serialize.url("pipeline_name", pipeline_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_request, 'ArtifactRenameRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _rename_pipeline_initial.metadata = {'url': '/pipelines/{pipelineName}/rename'} # type: ignore - - def begin_rename_pipeline( - self, - pipeline_name, # type: str - new_name=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> LROPoller[None] - """Renames a pipeline. - - :param pipeline_name: The pipeline name. - :type pipeline_name: str - :param new_name: New name of the artifact. - :type new_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._rename_pipeline_initial( - pipeline_name=pipeline_name, - new_name=new_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'pipelineName': self._serialize.url("pipeline_name", pipeline_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_rename_pipeline.metadata = {'url': '/pipelines/{pipelineName}/rename'} # type: ignore - - def create_pipeline_run( - self, - pipeline_name, # type: str - reference_pipeline_run_id=None, # type: Optional[str] - is_recovery=None, # type: Optional[bool] - start_activity_name=None, # type: Optional[str] - parameters=None, # type: Optional[Dict[str, Any]] - **kwargs # type: Any - ): - # type: (...) -> "_models.CreateRunResponse" - """Creates a run of a pipeline. - - :param pipeline_name: The pipeline name. - :type pipeline_name: str - :param reference_pipeline_run_id: The pipeline run identifier. If run ID is specified the - parameters of the specified run will be used to create a new run. - :type reference_pipeline_run_id: str - :param is_recovery: Recovery mode flag. If recovery mode is set to true, the specified - referenced pipeline run and the new run will be grouped under the same groupId. - :type is_recovery: bool - :param start_activity_name: In recovery mode, the rerun will start from this activity. If not - specified, all activities will run. - :type start_activity_name: str - :param parameters: Parameters of the pipeline run. These parameters will be used only if the - runId is not specified. - :type parameters: dict[str, any] - :keyword callable cls: A custom type or function that will be passed the direct response - :return: CreateRunResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.CreateRunResponse - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.CreateRunResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self.create_pipeline_run.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'pipelineName': self._serialize.url("pipeline_name", pipeline_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - if reference_pipeline_run_id is not None: - query_parameters['referencePipelineRunId'] = self._serialize.query("reference_pipeline_run_id", reference_pipeline_run_id, 'str') - if is_recovery is not None: - query_parameters['isRecovery'] = self._serialize.query("is_recovery", is_recovery, 'bool') - if start_activity_name is not None: - query_parameters['startActivityName'] = self._serialize.query("start_activity_name", start_activity_name, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - if parameters is not None: - body_content = self._serialize.body(parameters, '{object}') - else: - body_content = None - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('CreateRunResponse', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - create_pipeline_run.metadata = {'url': '/pipelines/{pipelineName}/createRun'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_pipeline_run_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_pipeline_run_operations.py deleted file mode 100644 index cdef6162e486..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_pipeline_run_operations.py +++ /dev/null @@ -1,287 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Optional, TypeVar - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class PipelineRunOperations(object): - """PipelineRunOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def query_pipeline_runs_by_workspace( - self, - filter_parameters, # type: "_models.RunFilterParameters" - **kwargs # type: Any - ): - # type: (...) -> "_models.PipelineRunsQueryResponse" - """Query pipeline runs in the workspace based on input filter conditions. - - :param filter_parameters: Parameters to filter the pipeline run. - :type filter_parameters: ~azure.synapse.artifacts.v2021_06_01_preview.models.RunFilterParameters - :keyword callable cls: A custom type or function that will be passed the direct response - :return: PipelineRunsQueryResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineRunsQueryResponse - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.PipelineRunsQueryResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self.query_pipeline_runs_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(filter_parameters, 'RunFilterParameters') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('PipelineRunsQueryResponse', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - query_pipeline_runs_by_workspace.metadata = {'url': '/queryPipelineRuns'} # type: ignore - - def get_pipeline_run( - self, - run_id, # type: str - **kwargs # type: Any - ): - # type: (...) -> "_models.PipelineRun" - """Get a pipeline run by its run ID. - - :param run_id: The pipeline run identifier. - :type run_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: PipelineRun, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.PipelineRun - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.PipelineRun"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_pipeline_run.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'runId': self._serialize.url("run_id", run_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('PipelineRun', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_pipeline_run.metadata = {'url': '/pipelineruns/{runId}'} # type: ignore - - def query_activity_runs( - self, - pipeline_name, # type: str - run_id, # type: str - filter_parameters, # type: "_models.RunFilterParameters" - **kwargs # type: Any - ): - # type: (...) -> "_models.ActivityRunsQueryResponse" - """Query activity runs based on input filter conditions. - - :param pipeline_name: The pipeline name. - :type pipeline_name: str - :param run_id: The pipeline run identifier. - :type run_id: str - :param filter_parameters: Parameters to filter the activity runs. - :type filter_parameters: ~azure.synapse.artifacts.v2021_06_01_preview.models.RunFilterParameters - :keyword callable cls: A custom type or function that will be passed the direct response - :return: ActivityRunsQueryResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.ActivityRunsQueryResponse - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.ActivityRunsQueryResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self.query_activity_runs.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'pipelineName': self._serialize.url("pipeline_name", pipeline_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - 'runId': self._serialize.url("run_id", run_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(filter_parameters, 'RunFilterParameters') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('ActivityRunsQueryResponse', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - query_activity_runs.metadata = {'url': '/pipelines/{pipelineName}/pipelineruns/{runId}/queryActivityruns'} # type: ignore - - def cancel_pipeline_run( - self, - run_id, # type: str - is_recursive=None, # type: Optional[bool] - **kwargs # type: Any - ): - # type: (...) -> None - """Cancel a pipeline run by its run ID. - - :param run_id: The pipeline run identifier. - :type run_id: str - :param is_recursive: If true, cancel all the Child pipelines that are triggered by the current - pipeline. - :type is_recursive: bool - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.cancel_pipeline_run.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'runId': self._serialize.url("run_id", run_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - if is_recursive is not None: - query_parameters['isRecursive'] = self._serialize.query("is_recursive", is_recursive, 'bool') - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - cancel_pipeline_run.metadata = {'url': '/pipelineruns/{runId}/cancel'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_spark_job_definition_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_spark_job_definition_operations.py deleted file mode 100644 index 52cf8316bb2e..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_spark_job_definition_operations.py +++ /dev/null @@ -1,769 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.paging import ItemPaged -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse -from azure.core.polling import LROPoller, NoPolling, PollingMethod -from azure.core.polling.base_polling import LROBasePolling - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Iterable, Optional, TypeVar, Union - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class SparkJobDefinitionOperations(object): - """SparkJobDefinitionOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get_spark_job_definitions_by_workspace( - self, - **kwargs # type: Any - ): - # type: (...) -> Iterable["_models.SparkJobDefinitionsListResponse"] - """Lists spark job definitions. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either SparkJobDefinitionsListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobDefinitionsListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.SparkJobDefinitionsListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.get_spark_job_definitions_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - def extract_data(pipeline_response): - deserialized = self._deserialize('SparkJobDefinitionsListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, iter(list_of_elem) - - def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return ItemPaged( - get_next, extract_data - ) - get_spark_job_definitions_by_workspace.metadata = {'url': '/sparkJobDefinitions'} # type: ignore - - def _create_or_update_spark_job_definition_initial( - self, - spark_job_definition_name, # type: str - properties, # type: "_models.SparkJobDefinition" - if_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.SparkJobDefinitionResource"] - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.SparkJobDefinitionResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _spark_job_definition = _models.SparkJobDefinitionResource(properties=properties) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._create_or_update_spark_job_definition_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sparkJobDefinitionName': self._serialize.url("spark_job_definition_name", spark_job_definition_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_match is not None: - header_parameters['If-Match'] = self._serialize.header("if_match", if_match, 'str') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_spark_job_definition, 'SparkJobDefinitionResource') - body_content_kwargs['content'] = body_content - request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('SparkJobDefinitionResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _create_or_update_spark_job_definition_initial.metadata = {'url': '/sparkJobDefinitions/{sparkJobDefinitionName}'} # type: ignore - - def begin_create_or_update_spark_job_definition( - self, - spark_job_definition_name, # type: str - properties, # type: "_models.SparkJobDefinition" - if_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> LROPoller["_models.SparkJobDefinitionResource"] - """Creates or updates a Spark Job Definition. - - :param spark_job_definition_name: The spark job definition name. - :type spark_job_definition_name: str - :param properties: Properties of spark job definition. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobDefinition - :param if_match: ETag of the Spark Job Definition entity. Should only be specified for update, - for which it should match existing entity or can be * for unconditional update. - :type if_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either SparkJobDefinitionResource or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobDefinitionResource] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.SparkJobDefinitionResource"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._create_or_update_spark_job_definition_initial( - spark_job_definition_name=spark_job_definition_name, - properties=properties, - if_match=if_match, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('SparkJobDefinitionResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sparkJobDefinitionName': self._serialize.url("spark_job_definition_name", spark_job_definition_name, 'str'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create_or_update_spark_job_definition.metadata = {'url': '/sparkJobDefinitions/{sparkJobDefinitionName}'} # type: ignore - - def get_spark_job_definition( - self, - spark_job_definition_name, # type: str - if_none_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.SparkJobDefinitionResource"] - """Gets a Spark Job Definition. - - :param spark_job_definition_name: The spark job definition name. - :type spark_job_definition_name: str - :param if_none_match: ETag of the Spark Job Definition entity. Should only be specified for - get. If the ETag matches the existing entity tag, or if * was provided, then no content will be - returned. - :type if_none_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: SparkJobDefinitionResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobDefinitionResource or None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.SparkJobDefinitionResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_spark_job_definition.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sparkJobDefinitionName': self._serialize.url("spark_job_definition_name", spark_job_definition_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_none_match is not None: - header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 304]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('SparkJobDefinitionResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_spark_job_definition.metadata = {'url': '/sparkJobDefinitions/{sparkJobDefinitionName}'} # type: ignore - - def _delete_spark_job_definition_initial( - self, - spark_job_definition_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._delete_spark_job_definition_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sparkJobDefinitionName': self._serialize.url("spark_job_definition_name", spark_job_definition_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.delete(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _delete_spark_job_definition_initial.metadata = {'url': '/sparkJobDefinitions/{sparkJobDefinitionName}'} # type: ignore - - def begin_delete_spark_job_definition( - self, - spark_job_definition_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> LROPoller[None] - """Deletes a Spark Job Definition. - - :param spark_job_definition_name: The spark job definition name. - :type spark_job_definition_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._delete_spark_job_definition_initial( - spark_job_definition_name=spark_job_definition_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sparkJobDefinitionName': self._serialize.url("spark_job_definition_name", spark_job_definition_name, 'str'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_delete_spark_job_definition.metadata = {'url': '/sparkJobDefinitions/{sparkJobDefinitionName}'} # type: ignore - - def _execute_spark_job_definition_initial( - self, - spark_job_definition_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> "_models.SparkBatchJob" - cls = kwargs.pop('cls', None) # type: ClsType["_models.SparkBatchJob"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._execute_spark_job_definition_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sparkJobDefinitionName': self._serialize.url("spark_job_definition_name", spark_job_definition_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if response.status_code == 200: - deserialized = self._deserialize('SparkBatchJob', pipeline_response) - - if response.status_code == 202: - deserialized = self._deserialize('SparkBatchJob', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _execute_spark_job_definition_initial.metadata = {'url': '/sparkJobDefinitions/{sparkJobDefinitionName}/execute'} # type: ignore - - def begin_execute_spark_job_definition( - self, - spark_job_definition_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> LROPoller["_models.SparkBatchJob"] - """Executes the spark job definition. - - :param spark_job_definition_name: The spark job definition name. - :type spark_job_definition_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either SparkBatchJob or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.SparkBatchJob] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.SparkBatchJob"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._execute_spark_job_definition_initial( - spark_job_definition_name=spark_job_definition_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('SparkBatchJob', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sparkJobDefinitionName': self._serialize.url("spark_job_definition_name", spark_job_definition_name, 'str'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, lro_options={'final-state-via': 'location'}, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_execute_spark_job_definition.metadata = {'url': '/sparkJobDefinitions/{sparkJobDefinitionName}/execute'} # type: ignore - - def _rename_spark_job_definition_initial( - self, - spark_job_definition_name, # type: str - new_name=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> None - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _request = _models.ArtifactRenameRequest(new_name=new_name) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._rename_spark_job_definition_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sparkJobDefinitionName': self._serialize.url("spark_job_definition_name", spark_job_definition_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_request, 'ArtifactRenameRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _rename_spark_job_definition_initial.metadata = {'url': '/sparkJobDefinitions/{sparkJobDefinitionName}/rename'} # type: ignore - - def begin_rename_spark_job_definition( - self, - spark_job_definition_name, # type: str - new_name=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> LROPoller[None] - """Renames a sparkJobDefinition. - - :param spark_job_definition_name: The spark job definition name. - :type spark_job_definition_name: str - :param new_name: New name of the artifact. - :type new_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._rename_spark_job_definition_initial( - spark_job_definition_name=spark_job_definition_name, - new_name=new_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sparkJobDefinitionName': self._serialize.url("spark_job_definition_name", spark_job_definition_name, 'str'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_rename_spark_job_definition.metadata = {'url': '/sparkJobDefinitions/{sparkJobDefinitionName}/rename'} # type: ignore - - def _debug_spark_job_definition_initial( - self, - properties, # type: "_models.SparkJobDefinition" - **kwargs # type: Any - ): - # type: (...) -> "_models.SparkBatchJob" - cls = kwargs.pop('cls', None) # type: ClsType["_models.SparkBatchJob"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _spark_job_definition_azure_resource = _models.SparkJobDefinitionResource(properties=properties) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._debug_spark_job_definition_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_spark_job_definition_azure_resource, 'SparkJobDefinitionResource') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if response.status_code == 200: - deserialized = self._deserialize('SparkBatchJob', pipeline_response) - - if response.status_code == 202: - deserialized = self._deserialize('SparkBatchJob', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _debug_spark_job_definition_initial.metadata = {'url': '/debugSparkJobDefinition'} # type: ignore - - def begin_debug_spark_job_definition( - self, - properties, # type: "_models.SparkJobDefinition" - **kwargs # type: Any - ): - # type: (...) -> LROPoller["_models.SparkBatchJob"] - """Debug the spark job definition. - - :param properties: Properties of spark job definition. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.SparkJobDefinition - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either SparkBatchJob or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.SparkBatchJob] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.SparkBatchJob"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._debug_spark_job_definition_initial( - properties=properties, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('SparkBatchJob', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, lro_options={'final-state-via': 'location'}, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_debug_spark_job_definition.metadata = {'url': '/debugSparkJobDefinition'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_sql_pools_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_sql_pools_operations.py deleted file mode 100644 index 701fd404a6cf..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_sql_pools_operations.py +++ /dev/null @@ -1,152 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Optional, TypeVar - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class SqlPoolsOperations(object): - """SqlPoolsOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def list( - self, - **kwargs # type: Any - ): - # type: (...) -> "_models.SqlPoolInfoListResult" - """List Sql Pools. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPoolInfoListResult, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPoolInfoListResult - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.SqlPoolInfoListResult"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.list.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('SqlPoolInfoListResult', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - list.metadata = {'url': '/sqlPools'} # type: ignore - - def get( - self, - sql_pool_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> "_models.SqlPool" - """Get Sql Pool. - - :param sql_pool_name: The Sql Pool name. - :type sql_pool_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlPool, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlPool - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.SqlPool"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sqlPoolName': self._serialize.url("sql_pool_name", sql_pool_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('SqlPool', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get.metadata = {'url': '/sqlPools/{sqlPoolName}'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_sql_script_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_sql_script_operations.py deleted file mode 100644 index efff3005766e..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_sql_script_operations.py +++ /dev/null @@ -1,533 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.paging import ItemPaged -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse -from azure.core.polling import LROPoller, NoPolling, PollingMethod -from azure.core.polling.base_polling import LROBasePolling - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Iterable, Optional, TypeVar, Union - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class SqlScriptOperations(object): - """SqlScriptOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get_sql_scripts_by_workspace( - self, - **kwargs # type: Any - ): - # type: (...) -> Iterable["_models.SqlScriptsListResponse"] - """Lists sql scripts. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either SqlScriptsListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.SqlScriptsListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.SqlScriptsListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.get_sql_scripts_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - def extract_data(pipeline_response): - deserialized = self._deserialize('SqlScriptsListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, iter(list_of_elem) - - def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return ItemPaged( - get_next, extract_data - ) - get_sql_scripts_by_workspace.metadata = {'url': '/sqlScripts'} # type: ignore - - def _create_or_update_sql_script_initial( - self, - sql_script_name, # type: str - sql_script, # type: "_models.SqlScriptResource" - if_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.SqlScriptResource"] - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.SqlScriptResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._create_or_update_sql_script_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sqlScriptName': self._serialize.url("sql_script_name", sql_script_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_match is not None: - header_parameters['If-Match'] = self._serialize.header("if_match", if_match, 'str') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(sql_script, 'SqlScriptResource') - body_content_kwargs['content'] = body_content - request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('SqlScriptResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _create_or_update_sql_script_initial.metadata = {'url': '/sqlScripts/{sqlScriptName}'} # type: ignore - - def begin_create_or_update_sql_script( - self, - sql_script_name, # type: str - sql_script, # type: "_models.SqlScriptResource" - if_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> LROPoller["_models.SqlScriptResource"] - """Creates or updates a Sql Script. - - :param sql_script_name: The sql script name. - :type sql_script_name: str - :param sql_script: Sql Script resource definition. - :type sql_script: ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlScriptResource - :param if_match: ETag of the SQL script entity. Should only be specified for update, for which - it should match existing entity or can be * for unconditional update. - :type if_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either SqlScriptResource or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.SqlScriptResource] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.SqlScriptResource"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._create_or_update_sql_script_initial( - sql_script_name=sql_script_name, - sql_script=sql_script, - if_match=if_match, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('SqlScriptResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sqlScriptName': self._serialize.url("sql_script_name", sql_script_name, 'str'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create_or_update_sql_script.metadata = {'url': '/sqlScripts/{sqlScriptName}'} # type: ignore - - def get_sql_script( - self, - sql_script_name, # type: str - if_none_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.SqlScriptResource"] - """Gets a sql script. - - :param sql_script_name: The sql script name. - :type sql_script_name: str - :param if_none_match: ETag of the sql compute entity. Should only be specified for get. If the - ETag matches the existing entity tag, or if * was provided, then no content will be returned. - :type if_none_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: SqlScriptResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.SqlScriptResource or None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.SqlScriptResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_sql_script.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sqlScriptName': self._serialize.url("sql_script_name", sql_script_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_none_match is not None: - header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 304]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('SqlScriptResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_sql_script.metadata = {'url': '/sqlScripts/{sqlScriptName}'} # type: ignore - - def _delete_sql_script_initial( - self, - sql_script_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._delete_sql_script_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sqlScriptName': self._serialize.url("sql_script_name", sql_script_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.delete(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _delete_sql_script_initial.metadata = {'url': '/sqlScripts/{sqlScriptName}'} # type: ignore - - def begin_delete_sql_script( - self, - sql_script_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> LROPoller[None] - """Deletes a Sql Script. - - :param sql_script_name: The sql script name. - :type sql_script_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._delete_sql_script_initial( - sql_script_name=sql_script_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sqlScriptName': self._serialize.url("sql_script_name", sql_script_name, 'str'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_delete_sql_script.metadata = {'url': '/sqlScripts/{sqlScriptName}'} # type: ignore - - def _rename_sql_script_initial( - self, - sql_script_name, # type: str - new_name=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> None - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _request = _models.ArtifactRenameRequest(new_name=new_name) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._rename_sql_script_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sqlScriptName': self._serialize.url("sql_script_name", sql_script_name, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_request, 'ArtifactRenameRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _rename_sql_script_initial.metadata = {'url': '/sqlScripts/{sqlScriptName}/rename'} # type: ignore - - def begin_rename_sql_script( - self, - sql_script_name, # type: str - new_name=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> LROPoller[None] - """Renames a sqlScript. - - :param sql_script_name: The sql script name. - :type sql_script_name: str - :param new_name: New name of the artifact. - :type new_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._rename_sql_script_initial( - sql_script_name=sql_script_name, - new_name=new_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'sqlScriptName': self._serialize.url("sql_script_name", sql_script_name, 'str'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_rename_sql_script.metadata = {'url': '/sqlScripts/{sqlScriptName}/rename'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_trigger_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_trigger_operations.py deleted file mode 100644 index 2408d03d5a70..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_trigger_operations.py +++ /dev/null @@ -1,907 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.paging import ItemPaged -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse -from azure.core.polling import LROPoller, NoPolling, PollingMethod -from azure.core.polling.base_polling import LROBasePolling - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Iterable, Optional, TypeVar, Union - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class TriggerOperations(object): - """TriggerOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get_triggers_by_workspace( - self, - **kwargs # type: Any - ): - # type: (...) -> Iterable["_models.TriggerListResponse"] - """Lists triggers. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: An iterator like instance of either TriggerListResponse or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerListResponse] - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.TriggerListResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - def prepare_request(next_link=None): - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - if not next_link: - # Construct URL - url = self.get_triggers_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - else: - url = next_link - query_parameters = {} # type: Dict[str, Any] - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - request = self._client.get(url, query_parameters, header_parameters) - return request - - def extract_data(pipeline_response): - deserialized = self._deserialize('TriggerListResponse', pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) - return deserialized.next_link or None, iter(list_of_elem) - - def get_next(next_link=None): - request = prepare_request(next_link) - - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, model=error) - - return pipeline_response - - return ItemPaged( - get_next, extract_data - ) - get_triggers_by_workspace.metadata = {'url': '/triggers'} # type: ignore - - def _create_or_update_trigger_initial( - self, - trigger_name, # type: str - properties, # type: "_models.Trigger" - if_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.TriggerResource"] - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.TriggerResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - - _trigger = _models.TriggerResource(properties=properties) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self._create_or_update_trigger_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_match is not None: - header_parameters['If-Match'] = self._serialize.header("if_match", if_match, 'str') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(_trigger, 'TriggerResource') - body_content_kwargs['content'] = body_content - request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('TriggerResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _create_or_update_trigger_initial.metadata = {'url': '/triggers/{triggerName}'} # type: ignore - - def begin_create_or_update_trigger( - self, - trigger_name, # type: str - properties, # type: "_models.Trigger" - if_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> LROPoller["_models.TriggerResource"] - """Creates or updates a trigger. - - :param trigger_name: The trigger name. - :type trigger_name: str - :param properties: Properties of the trigger. - :type properties: ~azure.synapse.artifacts.v2021_06_01_preview.models.Trigger - :param if_match: ETag of the trigger entity. Should only be specified for update, for which it - should match existing entity or can be * for unconditional update. - :type if_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either TriggerResource or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerResource] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.TriggerResource"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._create_or_update_trigger_initial( - trigger_name=trigger_name, - properties=properties, - if_match=if_match, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('TriggerResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_create_or_update_trigger.metadata = {'url': '/triggers/{triggerName}'} # type: ignore - - def get_trigger( - self, - trigger_name, # type: str - if_none_match=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.TriggerResource"] - """Gets a trigger. - - :param trigger_name: The trigger name. - :type trigger_name: str - :param if_none_match: ETag of the trigger entity. Should only be specified for get. If the ETag - matches the existing entity tag, or if * was provided, then no content will be returned. - :type if_none_match: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: TriggerResource, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerResource or None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.TriggerResource"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_trigger.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if if_none_match is not None: - header_parameters['If-None-Match'] = self._serialize.header("if_none_match", if_none_match, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 304]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('TriggerResource', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_trigger.metadata = {'url': '/triggers/{triggerName}'} # type: ignore - - def _delete_trigger_initial( - self, - trigger_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._delete_trigger_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.delete(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202, 204]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _delete_trigger_initial.metadata = {'url': '/triggers/{triggerName}'} # type: ignore - - def begin_delete_trigger( - self, - trigger_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> LROPoller[None] - """Deletes a trigger. - - :param trigger_name: The trigger name. - :type trigger_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._delete_trigger_initial( - trigger_name=trigger_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_delete_trigger.metadata = {'url': '/triggers/{triggerName}'} # type: ignore - - def _subscribe_trigger_to_events_initial( - self, - trigger_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.TriggerSubscriptionOperationStatus"] - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.TriggerSubscriptionOperationStatus"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._subscribe_trigger_to_events_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('TriggerSubscriptionOperationStatus', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _subscribe_trigger_to_events_initial.metadata = {'url': '/triggers/{triggerName}/subscribeToEvents'} # type: ignore - - def begin_subscribe_trigger_to_events( - self, - trigger_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> LROPoller["_models.TriggerSubscriptionOperationStatus"] - """Subscribe event trigger to events. - - :param trigger_name: The trigger name. - :type trigger_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either TriggerSubscriptionOperationStatus or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerSubscriptionOperationStatus] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.TriggerSubscriptionOperationStatus"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._subscribe_trigger_to_events_initial( - trigger_name=trigger_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('TriggerSubscriptionOperationStatus', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_subscribe_trigger_to_events.metadata = {'url': '/triggers/{triggerName}/subscribeToEvents'} # type: ignore - - def get_event_subscription_status( - self, - trigger_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> "_models.TriggerSubscriptionOperationStatus" - """Get a trigger's event subscription status. - - :param trigger_name: The trigger name. - :type trigger_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: TriggerSubscriptionOperationStatus, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerSubscriptionOperationStatus - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.TriggerSubscriptionOperationStatus"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get_event_subscription_status.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('TriggerSubscriptionOperationStatus', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_event_subscription_status.metadata = {'url': '/triggers/{triggerName}/getEventSubscriptionStatus'} # type: ignore - - def _unsubscribe_trigger_from_events_initial( - self, - trigger_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> Optional["_models.TriggerSubscriptionOperationStatus"] - cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.TriggerSubscriptionOperationStatus"]] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._unsubscribe_trigger_from_events_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200, 202]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = None - if response.status_code == 200: - deserialized = self._deserialize('TriggerSubscriptionOperationStatus', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - _unsubscribe_trigger_from_events_initial.metadata = {'url': '/triggers/{triggerName}/unsubscribeFromEvents'} # type: ignore - - def begin_unsubscribe_trigger_from_events( - self, - trigger_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> LROPoller["_models.TriggerSubscriptionOperationStatus"] - """Unsubscribe event trigger from events. - - :param trigger_name: The trigger name. - :type trigger_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either TriggerSubscriptionOperationStatus or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerSubscriptionOperationStatus] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType["_models.TriggerSubscriptionOperationStatus"] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._unsubscribe_trigger_from_events_initial( - trigger_name=trigger_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - deserialized = self._deserialize('TriggerSubscriptionOperationStatus', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - return deserialized - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_unsubscribe_trigger_from_events.metadata = {'url': '/triggers/{triggerName}/unsubscribeFromEvents'} # type: ignore - - def _start_trigger_initial( - self, - trigger_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._start_trigger_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _start_trigger_initial.metadata = {'url': '/triggers/{triggerName}/start'} # type: ignore - - def begin_start_trigger( - self, - trigger_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> LROPoller[None] - """Starts a trigger. - - :param trigger_name: The trigger name. - :type trigger_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._start_trigger_initial( - trigger_name=trigger_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_start_trigger.metadata = {'url': '/triggers/{triggerName}/start'} # type: ignore - - def _stop_trigger_initial( - self, - trigger_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self._stop_trigger_initial.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - _stop_trigger_initial.metadata = {'url': '/triggers/{triggerName}/stop'} # type: ignore - - def begin_stop_trigger( - self, - trigger_name, # type: str - **kwargs # type: Any - ): - # type: (...) -> LROPoller[None] - """Stops a trigger. - - :param trigger_name: The trigger name. - :type trigger_name: str - :keyword callable cls: A custom type or function that will be passed the direct response - :keyword str continuation_token: A continuation token to restart a poller from a saved state. - :keyword polling: By default, your polling method will be LROBasePolling. - Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. - :paramtype polling: bool or ~azure.core.polling.PollingMethod - :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. - :return: An instance of LROPoller that returns either None or the result of cls(response) - :rtype: ~azure.core.polling.LROPoller[None] - :raises ~azure.core.exceptions.HttpResponseError: - """ - polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] - cls = kwargs.pop('cls', None) # type: ClsType[None] - lro_delay = kwargs.pop( - 'polling_interval', - self._config.polling_interval - ) - cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] - if cont_token is None: - raw_result = self._stop_trigger_initial( - trigger_name=trigger_name, - cls=lambda x,y,z: x, - **kwargs - ) - - kwargs.pop('error_map', None) - kwargs.pop('content_type', None) - - def get_long_running_output(pipeline_response): - if cls: - return cls(pipeline_response, None, {}) - - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - } - - if polling is True: polling_method = LROBasePolling(lro_delay, path_format_arguments=path_format_arguments, **kwargs) - elif polling is False: polling_method = NoPolling() - else: polling_method = polling - if cont_token: - return LROPoller.from_continuation_token( - polling_method=polling_method, - continuation_token=cont_token, - client=self._client, - deserialization_callback=get_long_running_output - ) - else: - return LROPoller(self._client, raw_result, get_long_running_output, polling_method) - begin_stop_trigger.metadata = {'url': '/triggers/{triggerName}/stop'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_trigger_run_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_trigger_run_operations.py deleted file mode 100644 index e8acab989fe2..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_trigger_run_operations.py +++ /dev/null @@ -1,218 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Optional, TypeVar - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class TriggerRunOperations(object): - """TriggerRunOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def rerun_trigger_instance( - self, - trigger_name, # type: str - run_id, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - """Rerun single trigger instance by runId. - - :param trigger_name: The trigger name. - :type trigger_name: str - :param run_id: The pipeline run identifier. - :type run_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.rerun_trigger_instance.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - 'runId': self._serialize.url("run_id", run_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - rerun_trigger_instance.metadata = {'url': '/triggers/{triggerName}/triggerRuns/{runId}/rerun'} # type: ignore - - def cancel_trigger_instance( - self, - trigger_name, # type: str - run_id, # type: str - **kwargs # type: Any - ): - # type: (...) -> None - """Cancel single trigger instance by runId. - - :param trigger_name: The trigger name. - :type trigger_name: str - :param run_id: The pipeline run identifier. - :type run_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: None, or the result of cls(response) - :rtype: None - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType[None] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.cancel_trigger_instance.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - 'triggerName': self._serialize.url("trigger_name", trigger_name, 'str', max_length=260, min_length=1, pattern=r'^[A-Za-z0-9_][^<>*#.%&:\\+?/]*$'), - 'runId': self._serialize.url("run_id", run_id, 'str'), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.post(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - if cls: - return cls(pipeline_response, None, {}) - - cancel_trigger_instance.metadata = {'url': '/triggers/{triggerName}/triggerRuns/{runId}/cancel'} # type: ignore - - def query_trigger_runs_by_workspace( - self, - filter_parameters, # type: "_models.RunFilterParameters" - **kwargs # type: Any - ): - # type: (...) -> "_models.TriggerRunsQueryResponse" - """Query trigger runs. - - :param filter_parameters: Parameters to filter the pipeline run. - :type filter_parameters: ~azure.synapse.artifacts.v2021_06_01_preview.models.RunFilterParameters - :keyword callable cls: A custom type or function that will be passed the direct response - :return: TriggerRunsQueryResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.TriggerRunsQueryResponse - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.TriggerRunsQueryResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self.query_trigger_runs_by_workspace.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(filter_parameters, 'RunFilterParameters') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.CloudError, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('TriggerRunsQueryResponse', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - query_trigger_runs_by_workspace.metadata = {'url': '/queryTriggerRuns'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_workspace_git_repo_management_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_workspace_git_repo_management_operations.py deleted file mode 100644 index 0007ee152691..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_workspace_git_repo_management_operations.py +++ /dev/null @@ -1,109 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Optional, TypeVar - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class WorkspaceGitRepoManagementOperations(object): - """WorkspaceGitRepoManagementOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get_git_hub_access_token( - self, - git_hub_access_token_request, # type: "_models.GitHubAccessTokenRequest" - client_request_id=None, # type: Optional[str] - **kwargs # type: Any - ): - # type: (...) -> "_models.GitHubAccessTokenResponse" - """Get the GitHub access token. - - :param git_hub_access_token_request: - :type git_hub_access_token_request: ~azure.synapse.artifacts.v2021_06_01_preview.models.GitHubAccessTokenRequest - :param client_request_id: Can provide a guid, which is helpful for debugging and to provide - better customer support. - :type client_request_id: str - :keyword callable cls: A custom type or function that will be passed the direct response - :return: GitHubAccessTokenResponse, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.GitHubAccessTokenResponse - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.GitHubAccessTokenResponse"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - content_type = kwargs.pop("content_type", "application/json") - accept = "application/json" - - # Construct URL - url = self.get_git_hub_access_token.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - if client_request_id is not None: - header_parameters['x-ms-client-request-id'] = self._serialize.header("client_request_id", client_request_id, 'str') - header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - body_content_kwargs = {} # type: Dict[str, Any] - body_content = self._serialize.body(git_hub_access_token_request, 'GitHubAccessTokenRequest') - body_content_kwargs['content'] = body_content - request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response) - - deserialized = self._deserialize('GitHubAccessTokenResponse', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get_git_hub_access_token.metadata = {'url': '/getGitHubAccessToken'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_workspace_operations.py b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_workspace_operations.py deleted file mode 100644 index 28dd83d776ad..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/operations/_workspace_operations.py +++ /dev/null @@ -1,96 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- -from typing import TYPE_CHECKING -import warnings - -from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error -from azure.core.pipeline import PipelineResponse -from azure.core.pipeline.transport import HttpRequest, HttpResponse - -from .. import models as _models - -if TYPE_CHECKING: - # pylint: disable=unused-import,ungrouped-imports - from typing import Any, Callable, Dict, Generic, Optional, TypeVar - - T = TypeVar('T') - ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] - -class WorkspaceOperations(object): - """WorkspaceOperations operations. - - You should not instantiate this class directly. Instead, you should create a Client instance that - instantiates it for you and attaches it as an attribute. - - :ivar models: Alias to model classes used in this operation group. - :type models: ~azure.synapse.artifacts.v2021_06_01_preview.models - :param client: Client for service requests. - :param config: Configuration of service client. - :param serializer: An object model serializer. - :param deserializer: An object model deserializer. - """ - - models = _models - - def __init__(self, client, config, serializer, deserializer): - self._client = client - self._serialize = serializer - self._deserialize = deserializer - self._config = config - - def get( - self, - **kwargs # type: Any - ): - # type: (...) -> "_models.Workspace" - """Get Workspace. - - :keyword callable cls: A custom type or function that will be passed the direct response - :return: Workspace, or the result of cls(response) - :rtype: ~azure.synapse.artifacts.v2021_06_01_preview.models.Workspace - :raises: ~azure.core.exceptions.HttpResponseError - """ - cls = kwargs.pop('cls', None) # type: ClsType["_models.Workspace"] - error_map = { - 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError - } - error_map.update(kwargs.pop('error_map', {})) - api_version = "2021-06-01-preview" - accept = "application/json" - - # Construct URL - url = self.get.metadata['url'] # type: ignore - path_format_arguments = { - 'endpoint': self._serialize.url("self._config.endpoint", self._config.endpoint, 'str', skip_quote=True), - } - url = self._client.format_url(url, **path_format_arguments) - - # Construct parameters - query_parameters = {} # type: Dict[str, Any] - query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') - - # Construct headers - header_parameters = {} # type: Dict[str, Any] - header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') - - request = self._client.get(url, query_parameters, header_parameters) - pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorContract, response) - raise HttpResponseError(response=response, model=error) - - deserialized = self._deserialize('Workspace', pipeline_response) - - if cls: - return cls(pipeline_response, deserialized, {}) - - return deserialized - get.metadata = {'url': '/workspace'} # type: ignore diff --git a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/py.typed b/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/py.typed deleted file mode 100644 index e5aff4f83af8..000000000000 --- a/sdk/synapse/azure-synapse-artifacts/azure/synapse/artifacts/v2021_06_01_preview/py.typed +++ /dev/null @@ -1 +0,0 @@ -# Marker file for PEP 561. \ No newline at end of file diff --git a/sdk/synapse/azure-synapse-spark/CHANGELOG.md b/sdk/synapse/azure-synapse-spark/CHANGELOG.md index cd2e8ca2a95a..53607c1bed8d 100644 --- a/sdk/synapse/azure-synapse-spark/CHANGELOG.md +++ b/sdk/synapse/azure-synapse-spark/CHANGELOG.md @@ -1,14 +1,10 @@ # Release History -## 0.7.0 (Unreleased) +## 0.7.0 (2021-10-05) ### Features Added -### Breaking Changes - -### Bugs Fixed - -### Other Changes +- Added `LivyStatementStates` and `LivyStates` ## 0.6.0 (2021-08-10) diff --git a/sdk/synapse/azure-synapse-spark/azure/synapse/spark/models/__init__.py b/sdk/synapse/azure-synapse-spark/azure/synapse/spark/models/__init__.py index 692c34f27f23..b7d181f06684 100644 --- a/sdk/synapse/azure-synapse-spark/azure/synapse/spark/models/__init__.py +++ b/sdk/synapse/azure-synapse-spark/azure/synapse/spark/models/__init__.py @@ -44,6 +44,8 @@ from ._models import SparkStatementOutput # type: ignore from ._spark_client_enums import ( + LivyStatementStates, + LivyStates, PluginCurrentState, SchedulerCurrentState, SparkBatchJobResultType, @@ -71,6 +73,8 @@ 'SparkStatementCollection', 'SparkStatementOptions', 'SparkStatementOutput', + 'LivyStatementStates', + 'LivyStates', 'PluginCurrentState', 'SchedulerCurrentState', 'SparkBatchJobResultType', diff --git a/sdk/synapse/azure-synapse-spark/azure/synapse/spark/models/_models.py b/sdk/synapse/azure-synapse-spark/azure/synapse/spark/models/_models.py index c65549dca90e..4b0a18b5a9de 100644 --- a/sdk/synapse/azure-synapse-spark/azure/synapse/spark/models/_models.py +++ b/sdk/synapse/azure-synapse-spark/azure/synapse/spark/models/_models.py @@ -47,8 +47,9 @@ class SparkBatchJob(msrest.serialization.Model): :type app_id: str :param app_info: The detailed application info. :type app_info: dict[str, str] - :param state: The batch state. - :type state: str + :param state: The batch state. Possible values include: "not_started", "starting", "idle", + "busy", "shutting_down", "error", "dead", "killed", "success", "running", "recovering". + :type state: str or ~azure.synapse.spark.models.LivyStates :param log_lines: The log lines. :type log_lines: list[str] """ @@ -484,8 +485,9 @@ class SparkSession(msrest.serialization.Model): :type app_id: str :param app_info: Dictionary of :code:``. :type app_info: dict[str, str] - :param state: - :type state: str + :param state: The session state. Possible values include: "not_started", "starting", "idle", + "busy", "shutting_down", "error", "dead", "killed", "success", "running", "recovering". + :type state: str or ~azure.synapse.spark.models.LivyStates :param log_lines: :type log_lines: list[str] """ @@ -727,8 +729,9 @@ class SparkStatement(msrest.serialization.Model): :type id: int :param code: :type code: str - :param state: - :type state: str + :param state: Possible values include: "waiting", "running", "available", "error", + "cancelling", "cancelled". + :type state: str or ~azure.synapse.spark.models.LivyStatementStates :param output: :type output: ~azure.synapse.spark.models.SparkStatementOutput """ diff --git a/sdk/synapse/azure-synapse-spark/azure/synapse/spark/models/_models_py3.py b/sdk/synapse/azure-synapse-spark/azure/synapse/spark/models/_models_py3.py index 98f456d076ef..4ce13b395875 100644 --- a/sdk/synapse/azure-synapse-spark/azure/synapse/spark/models/_models_py3.py +++ b/sdk/synapse/azure-synapse-spark/azure/synapse/spark/models/_models_py3.py @@ -52,8 +52,9 @@ class SparkBatchJob(msrest.serialization.Model): :type app_id: str :param app_info: The detailed application info. :type app_info: dict[str, str] - :param state: The batch state. - :type state: str + :param state: The batch state. Possible values include: "not_started", "starting", "idle", + "busy", "shutting_down", "error", "dead", "killed", "success", "running", "recovering". + :type state: str or ~azure.synapse.spark.models.LivyStates :param log_lines: The log lines. :type log_lines: list[str] """ @@ -102,7 +103,7 @@ def __init__( tags: Optional[Dict[str, str]] = None, app_id: Optional[str] = None, app_info: Optional[Dict[str, str]] = None, - state: Optional[str] = None, + state: Optional[Union[str, "LivyStates"]] = None, log_lines: Optional[List[str]] = None, **kwargs ): @@ -571,8 +572,9 @@ class SparkSession(msrest.serialization.Model): :type app_id: str :param app_info: Dictionary of :code:``. :type app_info: dict[str, str] - :param state: - :type state: str + :param state: The session state. Possible values include: "not_started", "starting", "idle", + "busy", "shutting_down", "error", "dead", "killed", "success", "running", "recovering". + :type state: str or ~azure.synapse.spark.models.LivyStates :param log_lines: :type log_lines: list[str] """ @@ -621,7 +623,7 @@ def __init__( tags: Optional[Dict[str, str]] = None, app_id: Optional[str] = None, app_info: Optional[Dict[str, str]] = None, - state: Optional[str] = None, + state: Optional[Union[str, "LivyStates"]] = None, log_lines: Optional[List[str]] = None, **kwargs ): @@ -866,8 +868,9 @@ class SparkStatement(msrest.serialization.Model): :type id: int :param code: :type code: str - :param state: - :type state: str + :param state: Possible values include: "waiting", "running", "available", "error", + "cancelling", "cancelled". + :type state: str or ~azure.synapse.spark.models.LivyStatementStates :param output: :type output: ~azure.synapse.spark.models.SparkStatementOutput """ @@ -888,7 +891,7 @@ def __init__( *, id: int, code: Optional[str] = None, - state: Optional[str] = None, + state: Optional[Union[str, "LivyStatementStates"]] = None, output: Optional["SparkStatementOutput"] = None, **kwargs ): diff --git a/sdk/synapse/azure-synapse-spark/azure/synapse/spark/models/_spark_client_enums.py b/sdk/synapse/azure-synapse-spark/azure/synapse/spark/models/_spark_client_enums.py index a1312531d5dc..2089fb5f65ed 100644 --- a/sdk/synapse/azure-synapse-spark/azure/synapse/spark/models/_spark_client_enums.py +++ b/sdk/synapse/azure-synapse-spark/azure/synapse/spark/models/_spark_client_enums.py @@ -26,6 +26,31 @@ def __getattr__(cls, name): raise AttributeError(name) +class LivyStatementStates(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): + + WAITING = "waiting" + RUNNING = "running" + AVAILABLE = "available" + ERROR = "error" + CANCELLING = "cancelling" + CANCELLED = "cancelled" + +class LivyStates(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): + """The batch state + """ + + NOT_STARTED = "not_started" + STARTING = "starting" + IDLE = "idle" + BUSY = "busy" + SHUTTING_DOWN = "shutting_down" + ERROR = "error" + DEAD = "dead" + KILLED = "killed" + SUCCESS = "success" + RUNNING = "running" + RECOVERING = "recovering" + class PluginCurrentState(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): PREPARATION = "Preparation"