diff --git a/sdk/datafactory/azure-mgmt-datafactory/CHANGELOG.md b/sdk/datafactory/azure-mgmt-datafactory/CHANGELOG.md index a84b16da6e5e..76d0e0643e16 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/CHANGELOG.md +++ b/sdk/datafactory/azure-mgmt-datafactory/CHANGELOG.md @@ -1,5 +1,15 @@ # Release History +## 0.15.0 (2020-12-09) + +**Features** + + - Model Factory has a new parameter encryption + - Model ExecuteDataFlowActivity has a new parameter run_concurrently + - Model ExecuteDataFlowActivity has a new parameter trace_level + - Model ExecuteDataFlowActivity has a new parameter continue_on_error + - Model FactoryIdentity has a new parameter user_assigned_identities + ## 0.14.0 (2020-10-23) **Features** diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 7cc4b3b52518..c0aab6b04cea 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -123,6 +123,7 @@ from ._models_py3 import CassandraTableDataset from ._models_py3 import ChainingTrigger from ._models_py3 import CmdkeySetup + from ._models_py3 import CMKIdentityDefinition from ._models_py3 import CommonDataServiceForAppsEntityDataset from ._models_py3 import CommonDataServiceForAppsLinkedService from ._models_py3 import CommonDataServiceForAppsSink @@ -228,6 +229,7 @@ from ._models_py3 import EloquaLinkedService from ._models_py3 import EloquaObjectDataset from ._models_py3 import EloquaSource + from ._models_py3 import EncryptionConfiguration from ._models_py3 import EntityReference from ._models_py3 import EnvironmentVariableSetup from ._models_py3 import ExcelDataset @@ -753,6 +755,7 @@ from ._models import CassandraTableDataset from ._models import ChainingTrigger from ._models import CmdkeySetup + from ._models import CMKIdentityDefinition from ._models import CommonDataServiceForAppsEntityDataset from ._models import CommonDataServiceForAppsLinkedService from ._models import CommonDataServiceForAppsSink @@ -858,6 +861,7 @@ from ._models import EloquaLinkedService from ._models import EloquaObjectDataset from ._models import EloquaSource + from ._models import EncryptionConfiguration from ._models import EntityReference from ._models import EnvironmentVariableSetup from ._models import ExcelDataset @@ -1329,18 +1333,11 @@ DayOfWeek, DaysOfWeek, RecurrenceFrequency, - DataFlowComputeType, AzureFunctionActivityMethod, WebActivityMethod, - OraclePartitionOption, SalesforceSourceReadBehavior, - NetezzaPartitionOption, CassandraSourceReadConsistencyLevels, - TeradataPartitionOption, - SqlPartitionOption, StoredProcedureParameterType, - SapTablePartitionOption, - SapHanaPartitionOption, SsisPackageLocationType, HDInsightActivityDebugInfoOption, SalesforceSinkWriteBehavior, @@ -1359,9 +1356,16 @@ IntegrationRuntimeSsisCatalogPricingTier, IntegrationRuntimeLicenseType, IntegrationRuntimeEdition, + DataFlowComputeType, SsisObjectMetadataType, IntegrationRuntimeAuthKeyName, CopyBehaviorType, + SqlPartitionOption, + SapHanaPartitionOption, + SapTablePartitionOption, + OraclePartitionOption, + TeradataPartitionOption, + NetezzaPartitionOption, ) __all__ = [ @@ -1478,6 +1482,7 @@ 'CassandraTableDataset', 'ChainingTrigger', 'CmdkeySetup', + 'CMKIdentityDefinition', 'CommonDataServiceForAppsEntityDataset', 'CommonDataServiceForAppsLinkedService', 'CommonDataServiceForAppsSink', @@ -1583,6 +1588,7 @@ 'EloquaLinkedService', 'EloquaObjectDataset', 'EloquaSource', + 'EncryptionConfiguration', 'EntityReference', 'EnvironmentVariableSetup', 'ExcelDataset', @@ -2053,18 +2059,11 @@ 'DayOfWeek', 'DaysOfWeek', 'RecurrenceFrequency', - 'DataFlowComputeType', 'AzureFunctionActivityMethod', 'WebActivityMethod', - 'OraclePartitionOption', 'SalesforceSourceReadBehavior', - 'NetezzaPartitionOption', 'CassandraSourceReadConsistencyLevels', - 'TeradataPartitionOption', - 'SqlPartitionOption', 'StoredProcedureParameterType', - 'SapTablePartitionOption', - 'SapHanaPartitionOption', 'SsisPackageLocationType', 'HDInsightActivityDebugInfoOption', 'SalesforceSinkWriteBehavior', @@ -2083,7 +2082,14 @@ 'IntegrationRuntimeSsisCatalogPricingTier', 'IntegrationRuntimeLicenseType', 'IntegrationRuntimeEdition', + 'DataFlowComputeType', 'SsisObjectMetadataType', 'IntegrationRuntimeAuthKeyName', 'CopyBehaviorType', + 'SqlPartitionOption', + 'SapHanaPartitionOption', + 'SapTablePartitionOption', + 'OraclePartitionOption', + 'TeradataPartitionOption', + 'NetezzaPartitionOption', ] diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_data_factory_management_client_enums.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_data_factory_management_client_enums.py index 7ada06846cbb..da122b92c18c 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_data_factory_management_client_enums.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_data_factory_management_client_enums.py @@ -384,13 +384,6 @@ class RecurrenceFrequency(str, Enum): year = "Year" -class DataFlowComputeType(str, Enum): - - general = "General" - memory_optimized = "MemoryOptimized" - compute_optimized = "ComputeOptimized" - - class AzureFunctionActivityMethod(str, Enum): get = "GET" @@ -410,26 +403,12 @@ class WebActivityMethod(str, Enum): delete = "DELETE" -class OraclePartitionOption(str, Enum): - - none = "None" - physical_partitions_of_table = "PhysicalPartitionsOfTable" - dynamic_range = "DynamicRange" - - class SalesforceSourceReadBehavior(str, Enum): query = "Query" query_all = "QueryAll" -class NetezzaPartitionOption(str, Enum): - - none = "None" - data_slice = "DataSlice" - dynamic_range = "DynamicRange" - - class CassandraSourceReadConsistencyLevels(str, Enum): all = "ALL" @@ -444,20 +423,6 @@ class CassandraSourceReadConsistencyLevels(str, Enum): local_serial = "LOCAL_SERIAL" -class TeradataPartitionOption(str, Enum): - - none = "None" - hash = "Hash" - dynamic_range = "DynamicRange" - - -class SqlPartitionOption(str, Enum): - - none = "None" - physical_partitions_of_table = "PhysicalPartitionsOfTable" - dynamic_range = "DynamicRange" - - class StoredProcedureParameterType(str, Enum): string = "String" @@ -469,23 +434,6 @@ class StoredProcedureParameterType(str, Enum): date_enum = "Date" -class SapTablePartitionOption(str, Enum): - - none = "None" - partition_on_int = "PartitionOnInt" - partition_on_calendar_year = "PartitionOnCalendarYear" - partition_on_calendar_month = "PartitionOnCalendarMonth" - partition_on_calendar_date = "PartitionOnCalendarDate" - partition_on_time = "PartitionOnTime" - - -class SapHanaPartitionOption(str, Enum): - - none = "None" - physical_partitions_of_table = "PhysicalPartitionsOfTable" - sap_hana_dynamic_range = "SapHanaDynamicRange" - - class SsisPackageLocationType(str, Enum): ssisdb = "SSISDB" @@ -606,6 +554,13 @@ class IntegrationRuntimeEdition(str, Enum): enterprise = "Enterprise" +class DataFlowComputeType(str, Enum): + + general = "General" + memory_optimized = "MemoryOptimized" + compute_optimized = "ComputeOptimized" + + class SsisObjectMetadataType(str, Enum): folder = "Folder" @@ -625,3 +580,48 @@ class CopyBehaviorType(str, Enum): preserve_hierarchy = "PreserveHierarchy" flatten_hierarchy = "FlattenHierarchy" merge_files = "MergeFiles" + + +class SqlPartitionOption(str, Enum): + + none = "None" + physical_partitions_of_table = "PhysicalPartitionsOfTable" + dynamic_range = "DynamicRange" + + +class SapHanaPartitionOption(str, Enum): + + none = "None" + physical_partitions_of_table = "PhysicalPartitionsOfTable" + sap_hana_dynamic_range = "SapHanaDynamicRange" + + +class SapTablePartitionOption(str, Enum): + + none = "None" + partition_on_int = "PartitionOnInt" + partition_on_calendar_year = "PartitionOnCalendarYear" + partition_on_calendar_month = "PartitionOnCalendarMonth" + partition_on_calendar_date = "PartitionOnCalendarDate" + partition_on_time = "PartitionOnTime" + + +class OraclePartitionOption(str, Enum): + + none = "None" + physical_partitions_of_table = "PhysicalPartitionsOfTable" + dynamic_range = "DynamicRange" + + +class TeradataPartitionOption(str, Enum): + + none = "None" + hash = "Hash" + dynamic_range = "DynamicRange" + + +class NetezzaPartitionOption(str, Enum): + + none = "None" + data_slice = "DataSlice" + dynamic_range = "DynamicRange" diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models.py index 26c9212d4308..ebd28d0a36c0 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models.py @@ -6651,10 +6651,9 @@ class AzureSqlSource(TabularSource): :param produce_additional_types: Which additional types to produce. :type produce_additional_types: object :param partition_option: The partition mechanism that will be used for Sql - read in parallel. Possible values include: 'None', - 'PhysicalPartitionsOfTable', 'DynamicRange' - :type partition_option: str or - ~azure.mgmt.datafactory.models.SqlPartitionOption + read in parallel. Possible values include: "None", + "PhysicalPartitionsOfTable", "DynamicRange". + :type partition_option: object :param partition_settings: The settings that will be leveraged for Sql source partitioning. :type partition_settings: @@ -6677,7 +6676,7 @@ class AzureSqlSource(TabularSource): 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, 'produce_additional_types': {'key': 'produceAdditionalTypes', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, } @@ -8063,6 +8062,23 @@ def __init__(self, **kwargs): self.type = 'CmdkeySetup' +class CMKIdentityDefinition(Model): + """Managed Identity used for CMK. + + :param user_assigned_identity: The resource id of the user assigned + identity to authenticate to customer's key vault. + :type user_assigned_identity: str + """ + + _attribute_map = { + 'user_assigned_identity': {'key': 'userAssignedIdentity', 'type': 'str'}, + } + + def __init__(self, **kwargs): + super(CMKIdentityDefinition, self).__init__(**kwargs) + self.user_assigned_identity = kwargs.get('user_assigned_identity', None) + + class CommonDataServiceForAppsEntityDataset(Dataset): """The Common Data Service for Apps entity dataset. @@ -13200,6 +13216,46 @@ def __init__(self, **kwargs): self.type = 'EloquaSource' +class EncryptionConfiguration(Model): + """Definition of CMK for the factory. + + All required parameters must be populated in order to send to Azure. + + :param key_name: Required. The name of the key in Azure Key Vault to use + as Customer Managed Key. + :type key_name: str + :param vault_base_url: Required. The url of the Azure Key Vault used for + CMK. + :type vault_base_url: str + :param key_version: The version of the key used for CMK. If not provided, + latest version will be used. + :type key_version: str + :param identity: User assigned identity to use to authenticate to + customer's key vault. If not provided Managed Service Identity will be + used. + :type identity: ~azure.mgmt.datafactory.models.CMKIdentityDefinition + """ + + _validation = { + 'key_name': {'required': True}, + 'vault_base_url': {'required': True}, + } + + _attribute_map = { + 'key_name': {'key': 'keyName', 'type': 'str'}, + 'vault_base_url': {'key': 'vaultBaseUrl', 'type': 'str'}, + 'key_version': {'key': 'keyVersion', 'type': 'str'}, + 'identity': {'key': 'identity', 'type': 'CMKIdentityDefinition'}, + } + + def __init__(self, **kwargs): + super(EncryptionConfiguration, self).__init__(**kwargs) + self.key_name = kwargs.get('key_name', None) + self.vault_base_url = kwargs.get('vault_base_url', None) + self.key_version = kwargs.get('key_version', None) + self.identity = kwargs.get('identity', None) + + class EntityReference(Model): """The entity reference. @@ -13425,6 +13481,18 @@ class ExecuteDataFlowActivity(ExecutionActivity): :param compute: Compute properties for data flow activity. :type compute: ~azure.mgmt.datafactory.models.ExecuteDataFlowActivityTypePropertiesCompute + :param trace_level: Trace level setting used for data flow monitoring + output. Supported values are: 'coarse', 'fine', and 'none'. Type: string + (or Expression with resultType string) + :type trace_level: object + :param continue_on_error: Continue on error setting used for data flow + execution. Enables processing to continue if a sink fails. Type: boolean + (or Expression with resultType boolean) + :type continue_on_error: object + :param run_concurrently: Concurrent run setting used for data flow + execution. Allows sinks with the same save order to be processed + concurrently. Type: boolean (or Expression with resultType boolean) + :type run_concurrently: object """ _validation = { @@ -13446,6 +13514,9 @@ class ExecuteDataFlowActivity(ExecutionActivity): 'staging': {'key': 'typeProperties.staging', 'type': 'DataFlowStagingInfo'}, 'integration_runtime': {'key': 'typeProperties.integrationRuntime', 'type': 'IntegrationRuntimeReference'}, 'compute': {'key': 'typeProperties.compute', 'type': 'ExecuteDataFlowActivityTypePropertiesCompute'}, + 'trace_level': {'key': 'typeProperties.traceLevel', 'type': 'object'}, + 'continue_on_error': {'key': 'typeProperties.continueOnError', 'type': 'object'}, + 'run_concurrently': {'key': 'typeProperties.runConcurrently', 'type': 'object'}, } def __init__(self, **kwargs): @@ -13454,6 +13525,9 @@ def __init__(self, **kwargs): self.staging = kwargs.get('staging', None) self.integration_runtime = kwargs.get('integration_runtime', None) self.compute = kwargs.get('compute', None) + self.trace_level = kwargs.get('trace_level', None) + self.continue_on_error = kwargs.get('continue_on_error', None) + self.run_concurrently = kwargs.get('run_concurrently', None) self.type = 'ExecuteDataFlow' @@ -13462,17 +13536,17 @@ class ExecuteDataFlowActivityTypePropertiesCompute(Model): :param compute_type: Compute type of the cluster which will execute data flow job. Possible values include: 'General', 'MemoryOptimized', - 'ComputeOptimized' - :type compute_type: str or - ~azure.mgmt.datafactory.models.DataFlowComputeType + 'ComputeOptimized'. Type: string (or Expression with resultType string) + :type compute_type: object :param core_count: Core count of the cluster which will execute data flow - job. Supported values are: 8, 16, 32, 48, 80, 144 and 272. - :type core_count: int + job. Supported values are: 8, 16, 32, 48, 80, 144 and 272. Type: integer + (or Expression with resultType integer) + :type core_count: object """ _attribute_map = { - 'compute_type': {'key': 'computeType', 'type': 'str'}, - 'core_count': {'key': 'coreCount', 'type': 'int'}, + 'compute_type': {'key': 'computeType', 'type': 'object'}, + 'core_count': {'key': 'coreCount', 'type': 'object'}, } def __init__(self, **kwargs): @@ -13854,6 +13928,9 @@ class Factory(Resource): :param global_parameters: List of parameters for factory. :type global_parameters: dict[str, ~azure.mgmt.datafactory.models.GlobalParameterSpecification] + :param encryption: Properties to enable Customer Managed Key for the + factory. + :type encryption: ~azure.mgmt.datafactory.models.EncryptionConfiguration :param public_network_access: Whether or not public network access is allowed for the data factory. Possible values include: 'Enabled', 'Disabled' @@ -13885,6 +13962,7 @@ class Factory(Resource): 'version': {'key': 'properties.version', 'type': 'str'}, 'repo_configuration': {'key': 'properties.repoConfiguration', 'type': 'FactoryRepoConfiguration'}, 'global_parameters': {'key': 'properties.globalParameters', 'type': '{GlobalParameterSpecification}'}, + 'encryption': {'key': 'properties.encryption', 'type': 'EncryptionConfiguration'}, 'public_network_access': {'key': 'properties.publicNetworkAccess', 'type': 'str'}, } @@ -13897,6 +13975,7 @@ def __init__(self, **kwargs): self.version = None self.repo_configuration = kwargs.get('repo_configuration', None) self.global_parameters = kwargs.get('global_parameters', None) + self.encryption = kwargs.get('encryption', None) self.public_network_access = kwargs.get('public_network_access', None) @@ -14007,13 +14086,15 @@ class FactoryIdentity(Model): All required parameters must be populated in order to send to Azure. - :ivar type: Required. The identity type. Currently the only supported type - is 'SystemAssigned'. Default value: "SystemAssigned" . + :ivar type: Required. The identity type. Default value: "SystemAssigned" . :vartype type: str :ivar principal_id: The principal id of the identity. :vartype principal_id: str :ivar tenant_id: The client tenant id of the identity. :vartype tenant_id: str + :param user_assigned_identities: List of user assigned identities for the + factory. + :type user_assigned_identities: dict[str, object] """ _validation = { @@ -14026,6 +14107,7 @@ class FactoryIdentity(Model): 'type': {'key': 'type', 'type': 'str'}, 'principal_id': {'key': 'principalId', 'type': 'str'}, 'tenant_id': {'key': 'tenantId', 'type': 'str'}, + 'user_assigned_identities': {'key': 'userAssignedIdentities', 'type': '{object}'}, } type = "SystemAssigned" @@ -14034,6 +14116,7 @@ def __init__(self, **kwargs): super(FactoryIdentity, self).__init__(**kwargs) self.principal_id = None self.tenant_id = None + self.user_assigned_identities = kwargs.get('user_assigned_identities', None) class FactoryRepoUpdate(Model): @@ -22257,10 +22340,9 @@ class NetezzaSource(TabularSource): Expression with resultType string). :type query: object :param partition_option: The partition mechanism that will be used for - Netezza read in parallel. Possible values include: 'None', 'DataSlice', - 'DynamicRange' - :type partition_option: str or - ~azure.mgmt.datafactory.models.NetezzaPartitionOption + Netezza read in parallel. Possible values include: "None", "DataSlice", + "DynamicRange". + :type partition_option: object :param partition_settings: The settings that will be leveraged for Netezza source partitioning. :type partition_settings: @@ -22280,7 +22362,7 @@ class NetezzaSource(TabularSource): 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, 'query': {'key': 'query', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'NetezzaPartitionSettings'}, } @@ -23640,10 +23722,9 @@ class OracleSource(CopySource): ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: object :param partition_option: The partition mechanism that will be used for - Oracle read in parallel. Possible values include: 'None', - 'PhysicalPartitionsOfTable', 'DynamicRange' - :type partition_option: str or - ~azure.mgmt.datafactory.models.OraclePartitionOption + Oracle read in parallel. Possible values include: "None", + "PhysicalPartitionsOfTable", "DynamicRange". + :type partition_option: object :param partition_settings: The settings that will be leveraged for Oracle source partitioning. :type partition_settings: @@ -23667,7 +23748,7 @@ class OracleSource(CopySource): 'type': {'key': 'type', 'type': 'str'}, 'oracle_reader_query': {'key': 'oracleReaderQuery', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'OraclePartitionSettings'}, 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, } @@ -28173,10 +28254,9 @@ class SapHanaSource(TabularSource): integer(or Expression with resultType integer). :type packet_size: object :param partition_option: The partition mechanism that will be used for SAP - HANA read in parallel. Possible values include: 'None', - 'PhysicalPartitionsOfTable', 'SapHanaDynamicRange' - :type partition_option: str or - ~azure.mgmt.datafactory.models.SapHanaPartitionOption + HANA read in parallel. Possible values include: "None", + "PhysicalPartitionsOfTable", "SapHanaDynamicRange". + :type partition_option: object :param partition_settings: The settings that will be leveraged for SAP HANA source partitioning. :type partition_settings: @@ -28197,7 +28277,7 @@ class SapHanaSource(TabularSource): 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, 'query': {'key': 'query', 'type': 'object'}, 'packet_size': {'key': 'packetSize', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'SapHanaPartitionSettings'}, } @@ -28811,11 +28891,10 @@ class SapTableSource(TabularSource): retrieved. Type: string (or Expression with resultType string). :type sap_data_column_delimiter: object :param partition_option: The partition mechanism that will be used for SAP - table read in parallel. Possible values include: 'None', 'PartitionOnInt', - 'PartitionOnCalendarYear', 'PartitionOnCalendarMonth', - 'PartitionOnCalendarDate', 'PartitionOnTime' - :type partition_option: str or - ~azure.mgmt.datafactory.models.SapTablePartitionOption + table read in parallel. Possible values include: "None", "PartitionOnInt", + "PartitionOnCalendarYear", "PartitionOnCalendarMonth", + "PartitionOnCalendarDate", "PartitionOnTime". + :type partition_option: object :param partition_settings: The settings that will be leveraged for SAP table source partitioning. :type partition_settings: @@ -28841,7 +28920,7 @@ class SapTableSource(TabularSource): 'batch_size': {'key': 'batchSize', 'type': 'object'}, 'custom_rfc_read_table_function_module': {'key': 'customRfcReadTableFunctionModule', 'type': 'object'}, 'sap_data_column_delimiter': {'key': 'sapDataColumnDelimiter', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'SapTablePartitionSettings'}, } @@ -30960,10 +31039,9 @@ class SqlDWSource(TabularSource): StoredProcedureParameter. :type stored_procedure_parameters: object :param partition_option: The partition mechanism that will be used for Sql - read in parallel. Possible values include: 'None', - 'PhysicalPartitionsOfTable', 'DynamicRange' - :type partition_option: str or - ~azure.mgmt.datafactory.models.SqlPartitionOption + read in parallel. Possible values include: "None", + "PhysicalPartitionsOfTable", "DynamicRange". + :type partition_option: object :param partition_settings: The settings that will be leveraged for Sql source partitioning. :type partition_settings: @@ -30985,7 +31063,7 @@ class SqlDWSource(TabularSource): 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, } @@ -31124,10 +31202,9 @@ class SqlMISource(TabularSource): :param produce_additional_types: Which additional types to produce. :type produce_additional_types: object :param partition_option: The partition mechanism that will be used for Sql - read in parallel. Possible values include: 'None', - 'PhysicalPartitionsOfTable', 'DynamicRange' - :type partition_option: str or - ~azure.mgmt.datafactory.models.SqlPartitionOption + read in parallel. Possible values include: "None", + "PhysicalPartitionsOfTable", "DynamicRange". + :type partition_option: object :param partition_settings: The settings that will be leveraged for Sql source partitioning. :type partition_settings: @@ -31150,7 +31227,7 @@ class SqlMISource(TabularSource): 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, 'produce_additional_types': {'key': 'produceAdditionalTypes', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, } @@ -31387,10 +31464,9 @@ class SqlServerSource(TabularSource): :param produce_additional_types: Which additional types to produce. :type produce_additional_types: object :param partition_option: The partition mechanism that will be used for Sql - read in parallel. Possible values include: 'None', - 'PhysicalPartitionsOfTable', 'DynamicRange' - :type partition_option: str or - ~azure.mgmt.datafactory.models.SqlPartitionOption + read in parallel. Possible values include: "None", + "PhysicalPartitionsOfTable", "DynamicRange". + :type partition_option: object :param partition_settings: The settings that will be leveraged for Sql source partitioning. :type partition_settings: @@ -31413,7 +31489,7 @@ class SqlServerSource(TabularSource): 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, 'produce_additional_types': {'key': 'produceAdditionalTypes', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, } @@ -31685,10 +31761,9 @@ class SqlSource(TabularSource): resultType string). :type isolation_level: object :param partition_option: The partition mechanism that will be used for Sql - read in parallel. Possible values include: 'None', - 'PhysicalPartitionsOfTable', 'DynamicRange' - :type partition_option: str or - ~azure.mgmt.datafactory.models.SqlPartitionOption + read in parallel. Possible values include: "None", + "PhysicalPartitionsOfTable", "DynamicRange". + :type partition_option: object :param partition_settings: The settings that will be leveraged for Sql source partitioning. :type partition_settings: @@ -31711,7 +31786,7 @@ class SqlSource(TabularSource): 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, 'isolation_level': {'key': 'isolationLevel', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, } @@ -33182,10 +33257,9 @@ class TeradataSource(TabularSource): string). :type query: object :param partition_option: The partition mechanism that will be used for - teradata read in parallel. Possible values include: 'None', 'Hash', - 'DynamicRange' - :type partition_option: str or - ~azure.mgmt.datafactory.models.TeradataPartitionOption + teradata read in parallel. Possible values include: "None", "Hash", + "DynamicRange". + :type partition_option: object :param partition_settings: The settings that will be leveraged for teradata source partitioning. :type partition_settings: @@ -33205,7 +33279,7 @@ class TeradataSource(TabularSource): 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, 'query': {'key': 'query', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'TeradataPartitionSettings'}, } diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py index ece04cb31e05..39da9fd2c468 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py @@ -6651,10 +6651,9 @@ class AzureSqlSource(TabularSource): :param produce_additional_types: Which additional types to produce. :type produce_additional_types: object :param partition_option: The partition mechanism that will be used for Sql - read in parallel. Possible values include: 'None', - 'PhysicalPartitionsOfTable', 'DynamicRange' - :type partition_option: str or - ~azure.mgmt.datafactory.models.SqlPartitionOption + read in parallel. Possible values include: "None", + "PhysicalPartitionsOfTable", "DynamicRange". + :type partition_option: object :param partition_settings: The settings that will be leveraged for Sql source partitioning. :type partition_settings: @@ -6677,7 +6676,7 @@ class AzureSqlSource(TabularSource): 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, 'produce_additional_types': {'key': 'produceAdditionalTypes', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, } @@ -8063,6 +8062,23 @@ def __init__(self, *, target_name, user_name, password, **kwargs) -> None: self.type = 'CmdkeySetup' +class CMKIdentityDefinition(Model): + """Managed Identity used for CMK. + + :param user_assigned_identity: The resource id of the user assigned + identity to authenticate to customer's key vault. + :type user_assigned_identity: str + """ + + _attribute_map = { + 'user_assigned_identity': {'key': 'userAssignedIdentity', 'type': 'str'}, + } + + def __init__(self, *, user_assigned_identity: str=None, **kwargs) -> None: + super(CMKIdentityDefinition, self).__init__(**kwargs) + self.user_assigned_identity = user_assigned_identity + + class CommonDataServiceForAppsEntityDataset(Dataset): """The Common Data Service for Apps entity dataset. @@ -13200,6 +13216,46 @@ def __init__(self, *, additional_properties=None, source_retry_count=None, sourc self.type = 'EloquaSource' +class EncryptionConfiguration(Model): + """Definition of CMK for the factory. + + All required parameters must be populated in order to send to Azure. + + :param key_name: Required. The name of the key in Azure Key Vault to use + as Customer Managed Key. + :type key_name: str + :param vault_base_url: Required. The url of the Azure Key Vault used for + CMK. + :type vault_base_url: str + :param key_version: The version of the key used for CMK. If not provided, + latest version will be used. + :type key_version: str + :param identity: User assigned identity to use to authenticate to + customer's key vault. If not provided Managed Service Identity will be + used. + :type identity: ~azure.mgmt.datafactory.models.CMKIdentityDefinition + """ + + _validation = { + 'key_name': {'required': True}, + 'vault_base_url': {'required': True}, + } + + _attribute_map = { + 'key_name': {'key': 'keyName', 'type': 'str'}, + 'vault_base_url': {'key': 'vaultBaseUrl', 'type': 'str'}, + 'key_version': {'key': 'keyVersion', 'type': 'str'}, + 'identity': {'key': 'identity', 'type': 'CMKIdentityDefinition'}, + } + + def __init__(self, *, key_name: str, vault_base_url: str, key_version: str=None, identity=None, **kwargs) -> None: + super(EncryptionConfiguration, self).__init__(**kwargs) + self.key_name = key_name + self.vault_base_url = vault_base_url + self.key_version = key_version + self.identity = identity + + class EntityReference(Model): """The entity reference. @@ -13425,6 +13481,18 @@ class ExecuteDataFlowActivity(ExecutionActivity): :param compute: Compute properties for data flow activity. :type compute: ~azure.mgmt.datafactory.models.ExecuteDataFlowActivityTypePropertiesCompute + :param trace_level: Trace level setting used for data flow monitoring + output. Supported values are: 'coarse', 'fine', and 'none'. Type: string + (or Expression with resultType string) + :type trace_level: object + :param continue_on_error: Continue on error setting used for data flow + execution. Enables processing to continue if a sink fails. Type: boolean + (or Expression with resultType boolean) + :type continue_on_error: object + :param run_concurrently: Concurrent run setting used for data flow + execution. Allows sinks with the same save order to be processed + concurrently. Type: boolean (or Expression with resultType boolean) + :type run_concurrently: object """ _validation = { @@ -13446,14 +13514,20 @@ class ExecuteDataFlowActivity(ExecutionActivity): 'staging': {'key': 'typeProperties.staging', 'type': 'DataFlowStagingInfo'}, 'integration_runtime': {'key': 'typeProperties.integrationRuntime', 'type': 'IntegrationRuntimeReference'}, 'compute': {'key': 'typeProperties.compute', 'type': 'ExecuteDataFlowActivityTypePropertiesCompute'}, + 'trace_level': {'key': 'typeProperties.traceLevel', 'type': 'object'}, + 'continue_on_error': {'key': 'typeProperties.continueOnError', 'type': 'object'}, + 'run_concurrently': {'key': 'typeProperties.runConcurrently', 'type': 'object'}, } - def __init__(self, *, name: str, data_flow, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, staging=None, integration_runtime=None, compute=None, **kwargs) -> None: + def __init__(self, *, name: str, data_flow, additional_properties=None, description: str=None, depends_on=None, user_properties=None, linked_service_name=None, policy=None, staging=None, integration_runtime=None, compute=None, trace_level=None, continue_on_error=None, run_concurrently=None, **kwargs) -> None: super(ExecuteDataFlowActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.data_flow = data_flow self.staging = staging self.integration_runtime = integration_runtime self.compute = compute + self.trace_level = trace_level + self.continue_on_error = continue_on_error + self.run_concurrently = run_concurrently self.type = 'ExecuteDataFlow' @@ -13462,20 +13536,20 @@ class ExecuteDataFlowActivityTypePropertiesCompute(Model): :param compute_type: Compute type of the cluster which will execute data flow job. Possible values include: 'General', 'MemoryOptimized', - 'ComputeOptimized' - :type compute_type: str or - ~azure.mgmt.datafactory.models.DataFlowComputeType + 'ComputeOptimized'. Type: string (or Expression with resultType string) + :type compute_type: object :param core_count: Core count of the cluster which will execute data flow - job. Supported values are: 8, 16, 32, 48, 80, 144 and 272. - :type core_count: int + job. Supported values are: 8, 16, 32, 48, 80, 144 and 272. Type: integer + (or Expression with resultType integer) + :type core_count: object """ _attribute_map = { - 'compute_type': {'key': 'computeType', 'type': 'str'}, - 'core_count': {'key': 'coreCount', 'type': 'int'}, + 'compute_type': {'key': 'computeType', 'type': 'object'}, + 'core_count': {'key': 'coreCount', 'type': 'object'}, } - def __init__(self, *, compute_type=None, core_count: int=None, **kwargs) -> None: + def __init__(self, *, compute_type=None, core_count=None, **kwargs) -> None: super(ExecuteDataFlowActivityTypePropertiesCompute, self).__init__(**kwargs) self.compute_type = compute_type self.core_count = core_count @@ -13854,6 +13928,9 @@ class Factory(Resource): :param global_parameters: List of parameters for factory. :type global_parameters: dict[str, ~azure.mgmt.datafactory.models.GlobalParameterSpecification] + :param encryption: Properties to enable Customer Managed Key for the + factory. + :type encryption: ~azure.mgmt.datafactory.models.EncryptionConfiguration :param public_network_access: Whether or not public network access is allowed for the data factory. Possible values include: 'Enabled', 'Disabled' @@ -13885,10 +13962,11 @@ class Factory(Resource): 'version': {'key': 'properties.version', 'type': 'str'}, 'repo_configuration': {'key': 'properties.repoConfiguration', 'type': 'FactoryRepoConfiguration'}, 'global_parameters': {'key': 'properties.globalParameters', 'type': '{GlobalParameterSpecification}'}, + 'encryption': {'key': 'properties.encryption', 'type': 'EncryptionConfiguration'}, 'public_network_access': {'key': 'properties.publicNetworkAccess', 'type': 'str'}, } - def __init__(self, *, location: str=None, tags=None, additional_properties=None, identity=None, repo_configuration=None, global_parameters=None, public_network_access=None, **kwargs) -> None: + def __init__(self, *, location: str=None, tags=None, additional_properties=None, identity=None, repo_configuration=None, global_parameters=None, encryption=None, public_network_access=None, **kwargs) -> None: super(Factory, self).__init__(location=location, tags=tags, **kwargs) self.additional_properties = additional_properties self.identity = identity @@ -13897,6 +13975,7 @@ def __init__(self, *, location: str=None, tags=None, additional_properties=None, self.version = None self.repo_configuration = repo_configuration self.global_parameters = global_parameters + self.encryption = encryption self.public_network_access = public_network_access @@ -14007,13 +14086,15 @@ class FactoryIdentity(Model): All required parameters must be populated in order to send to Azure. - :ivar type: Required. The identity type. Currently the only supported type - is 'SystemAssigned'. Default value: "SystemAssigned" . + :ivar type: Required. The identity type. Default value: "SystemAssigned" . :vartype type: str :ivar principal_id: The principal id of the identity. :vartype principal_id: str :ivar tenant_id: The client tenant id of the identity. :vartype tenant_id: str + :param user_assigned_identities: List of user assigned identities for the + factory. + :type user_assigned_identities: dict[str, object] """ _validation = { @@ -14026,14 +14107,16 @@ class FactoryIdentity(Model): 'type': {'key': 'type', 'type': 'str'}, 'principal_id': {'key': 'principalId', 'type': 'str'}, 'tenant_id': {'key': 'tenantId', 'type': 'str'}, + 'user_assigned_identities': {'key': 'userAssignedIdentities', 'type': '{object}'}, } type = "SystemAssigned" - def __init__(self, **kwargs) -> None: + def __init__(self, *, user_assigned_identities=None, **kwargs) -> None: super(FactoryIdentity, self).__init__(**kwargs) self.principal_id = None self.tenant_id = None + self.user_assigned_identities = user_assigned_identities class FactoryRepoUpdate(Model): @@ -22257,10 +22340,9 @@ class NetezzaSource(TabularSource): Expression with resultType string). :type query: object :param partition_option: The partition mechanism that will be used for - Netezza read in parallel. Possible values include: 'None', 'DataSlice', - 'DynamicRange' - :type partition_option: str or - ~azure.mgmt.datafactory.models.NetezzaPartitionOption + Netezza read in parallel. Possible values include: "None", "DataSlice", + "DynamicRange". + :type partition_option: object :param partition_settings: The settings that will be leveraged for Netezza source partitioning. :type partition_settings: @@ -22280,7 +22362,7 @@ class NetezzaSource(TabularSource): 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, 'query': {'key': 'query', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'NetezzaPartitionSettings'}, } @@ -23640,10 +23722,9 @@ class OracleSource(CopySource): ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: object :param partition_option: The partition mechanism that will be used for - Oracle read in parallel. Possible values include: 'None', - 'PhysicalPartitionsOfTable', 'DynamicRange' - :type partition_option: str or - ~azure.mgmt.datafactory.models.OraclePartitionOption + Oracle read in parallel. Possible values include: "None", + "PhysicalPartitionsOfTable", "DynamicRange". + :type partition_option: object :param partition_settings: The settings that will be leveraged for Oracle source partitioning. :type partition_settings: @@ -23667,7 +23748,7 @@ class OracleSource(CopySource): 'type': {'key': 'type', 'type': 'str'}, 'oracle_reader_query': {'key': 'oracleReaderQuery', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'OraclePartitionSettings'}, 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, } @@ -28173,10 +28254,9 @@ class SapHanaSource(TabularSource): integer(or Expression with resultType integer). :type packet_size: object :param partition_option: The partition mechanism that will be used for SAP - HANA read in parallel. Possible values include: 'None', - 'PhysicalPartitionsOfTable', 'SapHanaDynamicRange' - :type partition_option: str or - ~azure.mgmt.datafactory.models.SapHanaPartitionOption + HANA read in parallel. Possible values include: "None", + "PhysicalPartitionsOfTable", "SapHanaDynamicRange". + :type partition_option: object :param partition_settings: The settings that will be leveraged for SAP HANA source partitioning. :type partition_settings: @@ -28197,7 +28277,7 @@ class SapHanaSource(TabularSource): 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, 'query': {'key': 'query', 'type': 'object'}, 'packet_size': {'key': 'packetSize', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'SapHanaPartitionSettings'}, } @@ -28811,11 +28891,10 @@ class SapTableSource(TabularSource): retrieved. Type: string (or Expression with resultType string). :type sap_data_column_delimiter: object :param partition_option: The partition mechanism that will be used for SAP - table read in parallel. Possible values include: 'None', 'PartitionOnInt', - 'PartitionOnCalendarYear', 'PartitionOnCalendarMonth', - 'PartitionOnCalendarDate', 'PartitionOnTime' - :type partition_option: str or - ~azure.mgmt.datafactory.models.SapTablePartitionOption + table read in parallel. Possible values include: "None", "PartitionOnInt", + "PartitionOnCalendarYear", "PartitionOnCalendarMonth", + "PartitionOnCalendarDate", "PartitionOnTime". + :type partition_option: object :param partition_settings: The settings that will be leveraged for SAP table source partitioning. :type partition_settings: @@ -28841,7 +28920,7 @@ class SapTableSource(TabularSource): 'batch_size': {'key': 'batchSize', 'type': 'object'}, 'custom_rfc_read_table_function_module': {'key': 'customRfcReadTableFunctionModule', 'type': 'object'}, 'sap_data_column_delimiter': {'key': 'sapDataColumnDelimiter', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'SapTablePartitionSettings'}, } @@ -30960,10 +31039,9 @@ class SqlDWSource(TabularSource): StoredProcedureParameter. :type stored_procedure_parameters: object :param partition_option: The partition mechanism that will be used for Sql - read in parallel. Possible values include: 'None', - 'PhysicalPartitionsOfTable', 'DynamicRange' - :type partition_option: str or - ~azure.mgmt.datafactory.models.SqlPartitionOption + read in parallel. Possible values include: "None", + "PhysicalPartitionsOfTable", "DynamicRange". + :type partition_option: object :param partition_settings: The settings that will be leveraged for Sql source partitioning. :type partition_settings: @@ -30985,7 +31063,7 @@ class SqlDWSource(TabularSource): 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, } @@ -31124,10 +31202,9 @@ class SqlMISource(TabularSource): :param produce_additional_types: Which additional types to produce. :type produce_additional_types: object :param partition_option: The partition mechanism that will be used for Sql - read in parallel. Possible values include: 'None', - 'PhysicalPartitionsOfTable', 'DynamicRange' - :type partition_option: str or - ~azure.mgmt.datafactory.models.SqlPartitionOption + read in parallel. Possible values include: "None", + "PhysicalPartitionsOfTable", "DynamicRange". + :type partition_option: object :param partition_settings: The settings that will be leveraged for Sql source partitioning. :type partition_settings: @@ -31150,7 +31227,7 @@ class SqlMISource(TabularSource): 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, 'produce_additional_types': {'key': 'produceAdditionalTypes', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, } @@ -31387,10 +31464,9 @@ class SqlServerSource(TabularSource): :param produce_additional_types: Which additional types to produce. :type produce_additional_types: object :param partition_option: The partition mechanism that will be used for Sql - read in parallel. Possible values include: 'None', - 'PhysicalPartitionsOfTable', 'DynamicRange' - :type partition_option: str or - ~azure.mgmt.datafactory.models.SqlPartitionOption + read in parallel. Possible values include: "None", + "PhysicalPartitionsOfTable", "DynamicRange". + :type partition_option: object :param partition_settings: The settings that will be leveraged for Sql source partitioning. :type partition_settings: @@ -31413,7 +31489,7 @@ class SqlServerSource(TabularSource): 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, 'produce_additional_types': {'key': 'produceAdditionalTypes', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, } @@ -31685,10 +31761,9 @@ class SqlSource(TabularSource): resultType string). :type isolation_level: object :param partition_option: The partition mechanism that will be used for Sql - read in parallel. Possible values include: 'None', - 'PhysicalPartitionsOfTable', 'DynamicRange' - :type partition_option: str or - ~azure.mgmt.datafactory.models.SqlPartitionOption + read in parallel. Possible values include: "None", + "PhysicalPartitionsOfTable", "DynamicRange". + :type partition_option: object :param partition_settings: The settings that will be leveraged for Sql source partitioning. :type partition_settings: @@ -31711,7 +31786,7 @@ class SqlSource(TabularSource): 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, 'isolation_level': {'key': 'isolationLevel', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'SqlPartitionSettings'}, } @@ -33182,10 +33257,9 @@ class TeradataSource(TabularSource): string). :type query: object :param partition_option: The partition mechanism that will be used for - teradata read in parallel. Possible values include: 'None', 'Hash', - 'DynamicRange' - :type partition_option: str or - ~azure.mgmt.datafactory.models.TeradataPartitionOption + teradata read in parallel. Possible values include: "None", "Hash", + "DynamicRange". + :type partition_option: object :param partition_settings: The settings that will be leveraged for teradata source partitioning. :type partition_settings: @@ -33205,7 +33279,7 @@ class TeradataSource(TabularSource): 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, 'query': {'key': 'query', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'str'}, + 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'TeradataPartitionSettings'}, } diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py index eba31a2798cf..2f147316d9cb 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py @@ -9,5 +9,5 @@ # regenerated. # -------------------------------------------------------------------------- -VERSION = "0.14.0" +VERSION = "0.15.0"