From 560584516aea04a1392f3fd770e52f637d2110bc Mon Sep 17 00:00:00 2001 From: SDKAuto Date: Mon, 3 Mar 2025 04:56:50 +0000 Subject: [PATCH] CodeGen from PR 3633 in test-repo-billy/azure-rest-api-specs Merge e264abbfa0513f2f1c0422ae562a86d9d58adb70 into 76bf8250a833bba98aa233f0d27fa900ce05aaa0 --- .../azure-mgmt-datafactory/_meta.json | 8 +- .../azure/mgmt/datafactory/_serialization.py | 16 +- .../azure/mgmt/datafactory/_version.py | 2 +- .../aio/operations/_factories_operations.py | 75 -- .../azure/mgmt/datafactory/models/__init__.py | 8 + .../_data_factory_management_client_enums.py | 12 + .../mgmt/datafactory/models/_models_py3.py | 701 ++++++++++++++++-- .../operations/_factories_operations.py | 98 --- .../generated_samples/factories_list.py | 40 - ...factory_management_factories_operations.py | 10 - ...y_management_factories_operations_async.py | 10 - .../azure-mgmt-datafactory/setup.py | 8 +- 12 files changed, 660 insertions(+), 328 deletions(-) delete mode 100644 sdk/datafactory/azure-mgmt-datafactory/generated_samples/factories_list.py diff --git a/sdk/datafactory/azure-mgmt-datafactory/_meta.json b/sdk/datafactory/azure-mgmt-datafactory/_meta.json index bf776bd01373..5771b4f5ed06 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/_meta.json +++ b/sdk/datafactory/azure-mgmt-datafactory/_meta.json @@ -1,11 +1,11 @@ { - "commit": "f06cffbda682a8cd225a8b16bc6f000d26d01612", - "repository_url": "https://github.com/Azure/azure-rest-api-specs", + "commit": "cb6cc29ff6237a582731478a1788985fdc79e32b", + "repository_url": "https://github.com/test-repo-billy/azure-rest-api-specs", "autorest": "3.10.2", "use": [ - "@autorest/python@6.26.4", + "@autorest/python@6.27.4", "@autorest/modelerfour@4.27.0" ], - "autorest_command": "autorest specification/datafactory/resource-manager/readme.md --generate-sample=True --generate-test=True --include-x-ms-examples-original-file=True --python --python-sdks-folder=/mnt/vss/_work/1/azure-sdk-for-python/sdk --use=@autorest/python@6.26.4 --use=@autorest/modelerfour@4.27.0 --version=3.10.2 --version-tolerant=False", + "autorest_command": "autorest specification/datafactory/resource-manager/readme.md --generate-sample=True --generate-test=True --include-x-ms-examples-original-file=True --python --python-sdks-folder=/mnt/vss/_work/1/s/azure-sdk-for-python/sdk --use=@autorest/python@6.27.4 --use=@autorest/modelerfour@4.27.0 --version=3.10.2 --version-tolerant=False", "readme": "specification/datafactory/resource-manager/readme.md" } \ No newline at end of file diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_serialization.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_serialization.py index ce17d1798ce7..b24ab2885450 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_serialization.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_serialization.py @@ -310,7 +310,7 @@ def _create_xml_node(tag, prefix=None, ns=None): return ET.Element(tag) -class Model(object): +class Model: """Mixin for all client request body/response body models to support serialization and deserialization. """ @@ -563,7 +563,7 @@ def _decode_attribute_map_key(key): return key.replace("\\.", ".") -class Serializer(object): # pylint: disable=too-many-public-methods +class Serializer: # pylint: disable=too-many-public-methods """Request object model serializer.""" basic_types = {str: "str", int: "int", bool: "bool", float: "float"} @@ -1441,7 +1441,7 @@ def xml_key_extractor(attr, attr_desc, data): # pylint: disable=unused-argument return children[0] -class Deserializer(object): +class Deserializer: """Response object model deserializer. :param dict classes: Class type dictionary for deserializing complex types. @@ -1683,17 +1683,21 @@ def _instantiate_model(self, response, attrs, additional_properties=None): subtype = getattr(response, "_subtype_map", {}) try: readonly = [ - k for k, v in response._validation.items() if v.get("readonly") # pylint: disable=protected-access + k + for k, v in response._validation.items() # pylint: disable=protected-access # type: ignore + if v.get("readonly") ] const = [ - k for k, v in response._validation.items() if v.get("constant") # pylint: disable=protected-access + k + for k, v in response._validation.items() # pylint: disable=protected-access # type: ignore + if v.get("constant") ] kwargs = {k: v for k, v in attrs.items() if k not in subtype and k not in readonly + const} response_obj = response(**kwargs) for attr in readonly: setattr(response_obj, attr, attrs.get(attr)) if additional_properties: - response_obj.additional_properties = additional_properties + response_obj.additional_properties = additional_properties # type: ignore return response_obj except TypeError as err: msg = "Unable to deserialize {} into model {}. ".format(kwargs, response) # type: ignore diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py index f89ed38360ab..c47f66669f1b 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py @@ -6,4 +6,4 @@ # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -VERSION = "9.1.0" +VERSION = "1.0.0" diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/aio/operations/_factories_operations.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/aio/operations/_factories_operations.py index 392de8b7be89..c7d331bea31a 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/aio/operations/_factories_operations.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/aio/operations/_factories_operations.py @@ -35,7 +35,6 @@ build_get_git_hub_access_token_request, build_get_request, build_list_by_resource_group_request, - build_list_request, build_update_request, ) @@ -66,80 +65,6 @@ def __init__(self, *args, **kwargs) -> None: self._serialize = input_args.pop(0) if input_args else kwargs.pop("serializer") self._deserialize = input_args.pop(0) if input_args else kwargs.pop("deserializer") - @distributed_trace - def list(self, **kwargs: Any) -> AsyncIterable["_models.Factory"]: - """Lists factories under the specified subscription. - - :return: An iterator like instance of either Factory or the result of cls(response) - :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.mgmt.datafactory.models.Factory] - :raises ~azure.core.exceptions.HttpResponseError: - """ - _headers = kwargs.pop("headers", {}) or {} - _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - - api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._config.api_version)) - cls: ClsType[_models.FactoryListResponse] = kwargs.pop("cls", None) - - error_map: MutableMapping = { - 401: ClientAuthenticationError, - 404: ResourceNotFoundError, - 409: ResourceExistsError, - 304: ResourceNotModifiedError, - } - error_map.update(kwargs.pop("error_map", {}) or {}) - - def prepare_request(next_link=None): - if not next_link: - - _request = build_list_request( - subscription_id=self._config.subscription_id, - api_version=api_version, - headers=_headers, - params=_params, - ) - _request.url = self._client.format_url(_request.url) - - else: - # make call to next link with the client's api-version - _parsed_next_link = urllib.parse.urlparse(next_link) - _next_request_params = case_insensitive_dict( - { - key: [urllib.parse.quote(v) for v in value] - for key, value in urllib.parse.parse_qs(_parsed_next_link.query).items() - } - ) - _next_request_params["api-version"] = self._config.api_version - _request = HttpRequest( - "GET", urllib.parse.urljoin(next_link, _parsed_next_link.path), params=_next_request_params - ) - _request.url = self._client.format_url(_request.url) - _request.method = "GET" - return _request - - async def extract_data(pipeline_response): - deserialized = self._deserialize("FactoryListResponse", pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) # type: ignore - return deserialized.next_link or None, AsyncList(list_of_elem) - - async def get_next(next_link=None): - _request = prepare_request(next_link) - - _stream = False - pipeline_response: PipelineResponse = await self._client._pipeline.run( # pylint: disable=protected-access - _request, stream=_stream, **kwargs - ) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, error_format=ARMErrorFormat) - - return pipeline_response - - return AsyncItemPaged(get_next, extract_data) - @overload async def configure_factory_repo( self, diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index ee44f781e28c..da831ea478b9 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -744,8 +744,10 @@ TarGZipReadSettings, TarReadSettings, TeamDeskLinkedService, + TeradataImportCommand, TeradataLinkedService, TeradataPartitionSettings, + TeradataSink, TeradataSource, TeradataTableDataset, TextFormat, @@ -847,6 +849,7 @@ GoogleAdWordsAuthenticationType, GoogleBigQueryAuthenticationType, GoogleBigQueryV2AuthenticationType, + GreenplumAuthenticationType, HBaseAuthenticationType, HDInsightActivityDebugInfoOption, HdiNodeTypes, @@ -877,6 +880,7 @@ NotebookReferenceType, ODataAadServicePrincipalCredentialType, ODataAuthenticationType, + OracleAuthenticationType, OraclePartitionOption, OrcCompressionCodec, ParameterType, @@ -1671,8 +1675,10 @@ "TarGZipReadSettings", "TarReadSettings", "TeamDeskLinkedService", + "TeradataImportCommand", "TeradataLinkedService", "TeradataPartitionSettings", + "TeradataSink", "TeradataSource", "TeradataTableDataset", "TextFormat", @@ -1771,6 +1777,7 @@ "GoogleAdWordsAuthenticationType", "GoogleBigQueryAuthenticationType", "GoogleBigQueryV2AuthenticationType", + "GreenplumAuthenticationType", "HBaseAuthenticationType", "HDInsightActivityDebugInfoOption", "HdiNodeTypes", @@ -1801,6 +1808,7 @@ "NotebookReferenceType", "ODataAadServicePrincipalCredentialType", "ODataAuthenticationType", + "OracleAuthenticationType", "OraclePartitionOption", "OrcCompressionCodec", "ParameterType", diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_data_factory_management_client_enums.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_data_factory_management_client_enums.py index ac24170e4c2a..7e8f6f3494cf 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_data_factory_management_client_enums.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_data_factory_management_client_enums.py @@ -377,6 +377,12 @@ class GoogleBigQueryV2AuthenticationType(str, Enum, metaclass=CaseInsensitiveEnu USER_AUTHENTICATION = "UserAuthentication" +class GreenplumAuthenticationType(str, Enum, metaclass=CaseInsensitiveEnumMeta): + """The authentication type to use. Type: string. Only used for V2.""" + + BASIC = "Basic" + + class HBaseAuthenticationType(str, Enum, metaclass=CaseInsensitiveEnumMeta): """The authentication mechanism to use to connect to the HBase server.""" @@ -620,6 +626,12 @@ class ODataAuthenticationType(str, Enum, metaclass=CaseInsensitiveEnumMeta): MANAGED_SERVICE_IDENTITY = "ManagedServiceIdentity" +class OracleAuthenticationType(str, Enum, metaclass=CaseInsensitiveEnumMeta): + """Authentication type for connecting to the Oracle database. Only used for Version 2.0.""" + + BASIC = "Basic" + + class OraclePartitionOption(str, Enum, metaclass=CaseInsensitiveEnumMeta): """The partition mechanism that will be used for Oracle read in parallel.""" diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py index c93dd92dc173..161838cd592d 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py @@ -601,7 +601,7 @@ class LinkedService(_serialization.Model): "Impala": "ImpalaLinkedService", "Informix": "InformixLinkedService", "Jira": "JiraLinkedService", - "LakeHouse": "LakeHouseLinkedService", + "Lakehouse": "LakeHouseLinkedService", "Magento": "MagentoLinkedService", "MariaDB": "MariaDBLinkedService", "Marketo": "MarketoLinkedService", @@ -992,7 +992,7 @@ class Dataset(_serialization.Model): "InformixTable": "InformixTableDataset", "JiraObject": "JiraObjectDataset", "Json": "JsonDataset", - "LakeHouseTable": "LakeHouseTableDataset", + "LakehouseTable": "LakeHouseTableDataset", "MagentoObject": "MagentoObjectDataset", "MariaDBTable": "MariaDBTableDataset", "MarketoObject": "MarketoObjectDataset", @@ -5185,7 +5185,7 @@ class CopySink(_serialization.Model): MongoDbV2Sink, OdbcSink, OracleSink, OrcSink, ParquetSink, RestSink, SalesforceServiceCloudSink, SalesforceServiceCloudV2Sink, SalesforceSink, SalesforceV2Sink, SapCloudForCustomerSink, SnowflakeSink, SnowflakeV2Sink, SqlDWSink, SqlMISink, SqlServerSink, - SqlSink, WarehouseSink + SqlSink, TeradataSink, WarehouseSink All required parameters must be populated in order to send to server. @@ -5275,6 +5275,7 @@ class CopySink(_serialization.Model): "SqlMISink": "SqlMISink", "SqlServerSink": "SqlServerSink", "SqlSink": "SqlSink", + "TeradataSink": "TeradataSink", "WarehouseSink": "WarehouseSink", } } @@ -5587,9 +5588,9 @@ class AvroWriteSettings(FormatWriteSettings): :ivar max_rows_per_file: Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). :vartype max_rows_per_file: JSON - :ivar file_name_prefix: Specifies the file name pattern - :code:``_:code:``.:code:`` when copy from non-file - based store without partitionOptions. Type: string (or Expression with resultType string). + :ivar file_name_prefix: Specifies the file name pattern :code:``_\\ + :code:``.\\ :code:`` when copy from non-file based store without + partitionOptions. Type: string (or Expression with resultType string). :vartype file_name_prefix: JSON """ @@ -5627,9 +5628,9 @@ def __init__( :keyword max_rows_per_file: Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). :paramtype max_rows_per_file: JSON - :keyword file_name_prefix: Specifies the file name pattern - :code:``_:code:``.:code:`` when copy from non-file - based store without partitionOptions. Type: string (or Expression with resultType string). + :keyword file_name_prefix: Specifies the file name pattern :code:``_\\ + :code:``.\\ :code:`` when copy from non-file based store without + partitionOptions. Type: string (or Expression with resultType string). :paramtype file_name_prefix: JSON """ super().__init__(additional_properties=additional_properties, **kwargs) @@ -7603,7 +7604,7 @@ class ImportSettings(_serialization.Model): """Import command settings. You probably want to use the sub-classes and not this class directly. Known sub-classes are: - AzureDatabricksDeltaLakeImportCommand, SnowflakeImportCopyCommand + AzureDatabricksDeltaLakeImportCommand, SnowflakeImportCopyCommand, TeradataImportCommand All required parameters must be populated in order to send to server. @@ -7627,6 +7628,7 @@ class ImportSettings(_serialization.Model): "type": { "AzureDatabricksDeltaLakeImportCommand": "AzureDatabricksDeltaLakeImportCommand", "SnowflakeImportCopyCommand": "SnowflakeImportCopyCommand", + "TeradataImportCommand": "TeradataImportCommand", } } @@ -8575,8 +8577,8 @@ class AzureDataExplorerLinkedService(LinkedService): :ivar annotations: List of tags that can be used for describing the linked service. :vartype annotations: list[JSON] :ivar endpoint: The endpoint of Azure Data Explorer (the engine's endpoint). URL will be in the - format https://:code:``.:code:``.kusto.windows.net. Type: string (or - Expression with resultType string). Required. + format https://\\ :code:``.\\ :code:``.kusto.windows.net. Type: string + (or Expression with resultType string). Required. :vartype endpoint: JSON :ivar service_principal_id: The ID of the service principal used to authenticate against Azure Data Explorer. Type: string (or Expression with resultType string). @@ -8648,8 +8650,8 @@ def __init__( :keyword annotations: List of tags that can be used for describing the linked service. :paramtype annotations: list[JSON] :keyword endpoint: The endpoint of Azure Data Explorer (the engine's endpoint). URL will be in - the format https://:code:``.:code:``.kusto.windows.net. Type: string - (or Expression with resultType string). Required. + the format https://\\ :code:``.\\ :code:``.kusto.windows.net. Type: + string (or Expression with resultType string). Required. :paramtype endpoint: JSON :keyword service_principal_id: The ID of the service principal used to authenticate against Azure Data Explorer. Type: string (or Expression with resultType string). @@ -10566,7 +10568,7 @@ class AzureFunctionLinkedService(LinkedService): :ivar annotations: List of tags that can be used for describing the linked service. :vartype annotations: list[JSON] :ivar function_app_url: The endpoint of the Azure Function App. URL will be in the format - https://:code:``.azurewebsites.net. Type: string (or Expression with resultType + https://\\ :code:``.azurewebsites.net. Type: string (or Expression with resultType string). Required. :vartype function_app_url: JSON :ivar function_key: Function or Host key for Azure Function App. @@ -10637,7 +10639,7 @@ def __init__( :keyword annotations: List of tags that can be used for describing the linked service. :paramtype annotations: list[JSON] :keyword function_app_url: The endpoint of the Azure Function App. URL will be in the format - https://:code:``.azurewebsites.net. Type: string (or Expression with resultType + https://\\ :code:``.azurewebsites.net. Type: string (or Expression with resultType string). Required. :paramtype function_app_url: JSON :keyword function_key: Function or Host key for Azure Function App. @@ -12425,6 +12427,33 @@ class AzurePostgreSqlLinkedService(LinkedService): :ivar encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string. :vartype encrypted_credential: str + :ivar service_principal_id: The ID of the service principal used to authenticate against Azure + Database for PostgreSQL Flexible server. Type: string (or Expression with resultType string). + :vartype service_principal_id: JSON + :ivar service_principal_key: The key of the service principal used to authenticate against + Azure Database for PostgreSQL Flexible server. + :vartype service_principal_key: ~azure.mgmt.datafactory.models.SecretBase + :ivar service_principal_credential_type: The service principal credential type to use in + Server-To-Server authentication. 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' + for certificate. Type: string (or Expression with resultType string). + :vartype service_principal_credential_type: JSON + :ivar service_principal_embedded_cert: Specify the base64 encoded certificate of your + application registered in Azure Active Directory. Type: string (or Expression with resultType + string). + :vartype service_principal_embedded_cert: ~azure.mgmt.datafactory.models.SecretBase + :ivar service_principal_embedded_cert_password: Specify the password of your certificate if + your certificate has a password and you are using AadServicePrincipal authentication. Type: + string (or Expression with resultType string). + :vartype service_principal_embedded_cert_password: ~azure.mgmt.datafactory.models.SecretBase + :ivar tenant: The name or ID of the tenant to which the service principal belongs. Type: string + (or Expression with resultType string). + :vartype tenant: JSON + :ivar azure_cloud_type: Indicates the azure cloud type of the service principle auth. Allowed + values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is the data + factory regions’ cloud type. Type: string (or Expression with resultType string). + :vartype azure_cloud_type: JSON + :ivar credential: The credential reference containing authentication information. + :vartype credential: ~azure.mgmt.datafactory.models.CredentialReference """ _validation = { @@ -12453,9 +12482,20 @@ class AzurePostgreSqlLinkedService(LinkedService): "encoding": {"key": "typeProperties.encoding", "type": "object"}, "password": {"key": "typeProperties.password", "type": "AzureKeyVaultSecretReference"}, "encrypted_credential": {"key": "typeProperties.encryptedCredential", "type": "str"}, + "service_principal_id": {"key": "typeProperties.servicePrincipalId", "type": "object"}, + "service_principal_key": {"key": "typeProperties.servicePrincipalKey", "type": "SecretBase"}, + "service_principal_credential_type": {"key": "typeProperties.servicePrincipalCredentialType", "type": "object"}, + "service_principal_embedded_cert": {"key": "typeProperties.servicePrincipalEmbeddedCert", "type": "SecretBase"}, + "service_principal_embedded_cert_password": { + "key": "typeProperties.servicePrincipalEmbeddedCertPassword", + "type": "SecretBase", + }, + "tenant": {"key": "typeProperties.tenant", "type": "object"}, + "azure_cloud_type": {"key": "typeProperties.azureCloudType", "type": "object"}, + "credential": {"key": "typeProperties.credential", "type": "CredentialReference"}, } - def __init__( + def __init__( # pylint: disable=too-many-locals self, *, additional_properties: Optional[Dict[str, JSON]] = None, @@ -12478,6 +12518,14 @@ def __init__( encoding: Optional[JSON] = None, password: Optional["_models.AzureKeyVaultSecretReference"] = None, encrypted_credential: Optional[str] = None, + service_principal_id: Optional[JSON] = None, + service_principal_key: Optional["_models.SecretBase"] = None, + service_principal_credential_type: Optional[JSON] = None, + service_principal_embedded_cert: Optional["_models.SecretBase"] = None, + service_principal_embedded_cert_password: Optional["_models.SecretBase"] = None, + tenant: Optional[JSON] = None, + azure_cloud_type: Optional[JSON] = None, + credential: Optional["_models.CredentialReference"] = None, **kwargs: Any ) -> None: """ @@ -12532,6 +12580,34 @@ def __init__( :keyword encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string. :paramtype encrypted_credential: str + :keyword service_principal_id: The ID of the service principal used to authenticate against + Azure Database for PostgreSQL Flexible server. Type: string (or Expression with resultType + string). + :paramtype service_principal_id: JSON + :keyword service_principal_key: The key of the service principal used to authenticate against + Azure Database for PostgreSQL Flexible server. + :paramtype service_principal_key: ~azure.mgmt.datafactory.models.SecretBase + :keyword service_principal_credential_type: The service principal credential type to use in + Server-To-Server authentication. 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' + for certificate. Type: string (or Expression with resultType string). + :paramtype service_principal_credential_type: JSON + :keyword service_principal_embedded_cert: Specify the base64 encoded certificate of your + application registered in Azure Active Directory. Type: string (or Expression with resultType + string). + :paramtype service_principal_embedded_cert: ~azure.mgmt.datafactory.models.SecretBase + :keyword service_principal_embedded_cert_password: Specify the password of your certificate if + your certificate has a password and you are using AadServicePrincipal authentication. Type: + string (or Expression with resultType string). + :paramtype service_principal_embedded_cert_password: ~azure.mgmt.datafactory.models.SecretBase + :keyword tenant: The name or ID of the tenant to which the service principal belongs. Type: + string (or Expression with resultType string). + :paramtype tenant: JSON + :keyword azure_cloud_type: Indicates the azure cloud type of the service principle auth. + Allowed values are AzurePublic, AzureChina, AzureUsGovernment, AzureGermany. Default value is + the data factory regions’ cloud type. Type: string (or Expression with resultType string). + :paramtype azure_cloud_type: JSON + :keyword credential: The credential reference containing authentication information. + :paramtype credential: ~azure.mgmt.datafactory.models.CredentialReference """ super().__init__( additional_properties=additional_properties, @@ -12557,6 +12633,14 @@ def __init__( self.encoding = encoding self.password = password self.encrypted_credential = encrypted_credential + self.service_principal_id = service_principal_id + self.service_principal_key = service_principal_key + self.service_principal_credential_type = service_principal_credential_type + self.service_principal_embedded_cert = service_principal_embedded_cert + self.service_principal_embedded_cert_password = service_principal_embedded_cert_password + self.tenant = tenant + self.azure_cloud_type = azure_cloud_type + self.credential = credential class AzurePostgreSqlSink(CopySink): @@ -16485,8 +16569,8 @@ class AzureSynapseArtifactsLinkedService(LinkedService): :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] :ivar annotations: List of tags that can be used for describing the linked service. :vartype annotations: list[JSON] - :ivar endpoint: https://:code:``.dev.azuresynapse.net, Azure Synapse Analytics - workspace URL. Type: string (or Expression with resultType string). Required. + :ivar endpoint: https://\\ :code:``.dev.azuresynapse.net, Azure Synapse + Analytics workspace URL. Type: string (or Expression with resultType string). Required. :vartype endpoint: JSON :ivar authentication: Required to specify MSI, if using system assigned managed identity as authentication method. Type: string (or Expression with resultType string). @@ -16543,7 +16627,7 @@ def __init__( :paramtype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] :keyword annotations: List of tags that can be used for describing the linked service. :paramtype annotations: list[JSON] - :keyword endpoint: https://:code:``.dev.azuresynapse.net, Azure Synapse + :keyword endpoint: https://\\ :code:``.dev.azuresynapse.net, Azure Synapse Analytics workspace URL. Type: string (or Expression with resultType string). Required. :paramtype endpoint: JSON :keyword authentication: Required to specify MSI, if using system assigned managed identity as @@ -25271,9 +25355,9 @@ class DelimitedTextWriteSettings(FormatWriteSettings): :ivar max_rows_per_file: Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). :vartype max_rows_per_file: JSON - :ivar file_name_prefix: Specifies the file name pattern - :code:``_:code:``.:code:`` when copy from non-file - based store without partitionOptions. Type: string (or Expression with resultType string). + :ivar file_name_prefix: Specifies the file name pattern :code:``_\\ + :code:``.\\ :code:`` when copy from non-file based store without + partitionOptions. Type: string (or Expression with resultType string). :vartype file_name_prefix: JSON """ @@ -25314,9 +25398,9 @@ def __init__( :keyword max_rows_per_file: Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). :paramtype max_rows_per_file: JSON - :keyword file_name_prefix: Specifies the file name pattern - :code:``_:code:``.:code:`` when copy from non-file - based store without partitionOptions. Type: string (or Expression with resultType string). + :keyword file_name_prefix: Specifies the file name pattern :code:``_\\ + :code:``.\\ :code:`` when copy from non-file based store without + partitionOptions. Type: string (or Expression with resultType string). :paramtype file_name_prefix: JSON """ super().__init__(additional_properties=additional_properties, **kwargs) @@ -33622,6 +33706,30 @@ class GreenplumLinkedService(LinkedService): :ivar encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string. :vartype encrypted_credential: str + :ivar authentication_type: The authentication type to use. Type: string. Only used for V2. + "Basic" + :vartype authentication_type: str or ~azure.mgmt.datafactory.models.GreenplumAuthenticationType + :ivar host: Host name for connection. Type: string. Only used for V2. + :vartype host: JSON + :ivar port: The port for the connection. Type: integer. Only used for V2. + :vartype port: JSON + :ivar username: Username for authentication. Type: string. Only used for V2. + :vartype username: JSON + :ivar database: Database name for connection. Type: string. Only used for V2. + :vartype database: JSON + :ivar ssl_mode: SSL mode for connection. Type: integer. 0: disable, 1:allow, 2: prefer, 3: + require, 4: verify-ca, 5: verify-full. Type: integer. Only used for V2. + :vartype ssl_mode: JSON + :ivar connection_timeout: The time to wait (in seconds) while trying to establish a connection + before terminating the attempt and generating an error. Type: integer. Only used for V2. + :vartype connection_timeout: JSON + :ivar command_timeout: The time to wait (in seconds) while trying to execute a command before + terminating the attempt and generating an error. Set to zero for infinity. Type: integer. Only + used for V2. + :vartype command_timeout: JSON + :ivar password: The Azure key vault secret reference of password in connection string. Type: + string. Only used for V2. + :vartype password: ~azure.mgmt.datafactory.models.SecretBase """ _validation = { @@ -33639,6 +33747,15 @@ class GreenplumLinkedService(LinkedService): "connection_string": {"key": "typeProperties.connectionString", "type": "object"}, "pwd": {"key": "typeProperties.pwd", "type": "AzureKeyVaultSecretReference"}, "encrypted_credential": {"key": "typeProperties.encryptedCredential", "type": "str"}, + "authentication_type": {"key": "typeProperties.authenticationType", "type": "str"}, + "host": {"key": "typeProperties.host", "type": "object"}, + "port": {"key": "typeProperties.port", "type": "object"}, + "username": {"key": "typeProperties.username", "type": "object"}, + "database": {"key": "typeProperties.database", "type": "object"}, + "ssl_mode": {"key": "typeProperties.sslMode", "type": "object"}, + "connection_timeout": {"key": "typeProperties.connectionTimeout", "type": "object"}, + "command_timeout": {"key": "typeProperties.commandTimeout", "type": "object"}, + "password": {"key": "typeProperties.password", "type": "SecretBase"}, } def __init__( @@ -33653,6 +33770,15 @@ def __init__( connection_string: Optional[JSON] = None, pwd: Optional["_models.AzureKeyVaultSecretReference"] = None, encrypted_credential: Optional[str] = None, + authentication_type: Optional[Union[str, "_models.GreenplumAuthenticationType"]] = None, + host: Optional[JSON] = None, + port: Optional[JSON] = None, + username: Optional[JSON] = None, + database: Optional[JSON] = None, + ssl_mode: Optional[JSON] = None, + connection_timeout: Optional[JSON] = None, + command_timeout: Optional[JSON] = None, + password: Optional["_models.SecretBase"] = None, **kwargs: Any ) -> None: """ @@ -33677,6 +33803,32 @@ def __init__( :keyword encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string. :paramtype encrypted_credential: str + :keyword authentication_type: The authentication type to use. Type: string. Only used for V2. + "Basic" + :paramtype authentication_type: str or + ~azure.mgmt.datafactory.models.GreenplumAuthenticationType + :keyword host: Host name for connection. Type: string. Only used for V2. + :paramtype host: JSON + :keyword port: The port for the connection. Type: integer. Only used for V2. + :paramtype port: JSON + :keyword username: Username for authentication. Type: string. Only used for V2. + :paramtype username: JSON + :keyword database: Database name for connection. Type: string. Only used for V2. + :paramtype database: JSON + :keyword ssl_mode: SSL mode for connection. Type: integer. 0: disable, 1:allow, 2: prefer, 3: + require, 4: verify-ca, 5: verify-full. Type: integer. Only used for V2. + :paramtype ssl_mode: JSON + :keyword connection_timeout: The time to wait (in seconds) while trying to establish a + connection before terminating the attempt and generating an error. Type: integer. Only used for + V2. + :paramtype connection_timeout: JSON + :keyword command_timeout: The time to wait (in seconds) while trying to execute a command + before terminating the attempt and generating an error. Set to zero for infinity. Type: + integer. Only used for V2. + :paramtype command_timeout: JSON + :keyword password: The Azure key vault secret reference of password in connection string. Type: + string. Only used for V2. + :paramtype password: ~azure.mgmt.datafactory.models.SecretBase """ super().__init__( additional_properties=additional_properties, @@ -33691,6 +33843,15 @@ def __init__( self.connection_string = connection_string self.pwd = pwd self.encrypted_credential = encrypted_credential + self.authentication_type = authentication_type + self.host = host + self.port = port + self.username = username + self.database = database + self.ssl_mode = ssl_mode + self.connection_timeout = connection_timeout + self.command_timeout = command_timeout + self.password = password class GreenplumSource(TabularSource): @@ -40738,7 +40899,7 @@ def __init__( class LakeHouseLinkedService(LinkedService): - """Microsoft Fabric LakeHouse linked service. + """Microsoft Fabric Lakehouse linked service. All required parameters must be populated in order to send to server. @@ -40760,14 +40921,14 @@ class LakeHouseLinkedService(LinkedService): :ivar workspace_id: The ID of Microsoft Fabric workspace. Type: string (or Expression with resultType string). :vartype workspace_id: JSON - :ivar artifact_id: The ID of Microsoft Fabric LakeHouse artifact. Type: string (or Expression + :ivar artifact_id: The ID of Microsoft Fabric Lakehouse artifact. Type: string (or Expression with resultType string). :vartype artifact_id: JSON :ivar service_principal_id: The ID of the application used to authenticate against Microsoft - Fabric LakeHouse. Type: string (or Expression with resultType string). + Fabric Lakehouse. Type: string (or Expression with resultType string). :vartype service_principal_id: JSON :ivar service_principal_key: The Key of the application used to authenticate against Microsoft - Fabric LakeHouse. + Fabric Lakehouse. :vartype service_principal_key: ~azure.mgmt.datafactory.models.SecretBase :ivar tenant: The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). @@ -40845,14 +41006,14 @@ def __init__( :keyword workspace_id: The ID of Microsoft Fabric workspace. Type: string (or Expression with resultType string). :paramtype workspace_id: JSON - :keyword artifact_id: The ID of Microsoft Fabric LakeHouse artifact. Type: string (or + :keyword artifact_id: The ID of Microsoft Fabric Lakehouse artifact. Type: string (or Expression with resultType string). :paramtype artifact_id: JSON :keyword service_principal_id: The ID of the application used to authenticate against Microsoft - Fabric LakeHouse. Type: string (or Expression with resultType string). + Fabric Lakehouse. Type: string (or Expression with resultType string). :paramtype service_principal_id: JSON :keyword service_principal_key: The Key of the application used to authenticate against - Microsoft Fabric LakeHouse. + Microsoft Fabric Lakehouse. :paramtype service_principal_key: ~azure.mgmt.datafactory.models.SecretBase :keyword tenant: The name or ID of the tenant to which the service principal belongs. Type: string (or Expression with resultType string). @@ -40880,7 +41041,7 @@ def __init__( annotations=annotations, **kwargs ) - self.type: str = "LakeHouse" + self.type: str = "Lakehouse" self.workspace_id = workspace_id self.artifact_id = artifact_id self.service_principal_id = service_principal_id @@ -40892,7 +41053,7 @@ def __init__( class LakeHouseLocation(DatasetLocation): - """The location of Microsoft Fabric LakeHouse Files dataset. + """The location of Microsoft Fabric Lakehouse Files dataset. All required parameters must be populated in order to send to server. @@ -40946,7 +41107,7 @@ def __init__( class LakeHouseReadSettings(StoreReadSettings): - """Microsoft Fabric LakeHouse Files read settings. + """Microsoft Fabric Lakehouse Files read settings. All required parameters must be populated in order to send to server. @@ -40964,10 +41125,10 @@ class LakeHouseReadSettings(StoreReadSettings): :ivar recursive: If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). :vartype recursive: JSON - :ivar wildcard_folder_path: Microsoft Fabric LakeHouse Files wildcardFolderPath. Type: string + :ivar wildcard_folder_path: Microsoft Fabric Lakehouse Files wildcardFolderPath. Type: string (or Expression with resultType string). :vartype wildcard_folder_path: JSON - :ivar wildcard_file_name: Microsoft Fabric LakeHouse Files wildcardFileName. Type: string (or + :ivar wildcard_file_name: Microsoft Fabric Lakehouse Files wildcardFileName. Type: string (or Expression with resultType string). :vartype wildcard_file_name: JSON :ivar file_list_path: Point to a text file that lists each file (relative path to the path @@ -41041,10 +41202,10 @@ def __init__( :keyword recursive: If true, files under the folder path will be read recursively. Default is true. Type: boolean (or Expression with resultType boolean). :paramtype recursive: JSON - :keyword wildcard_folder_path: Microsoft Fabric LakeHouse Files wildcardFolderPath. Type: + :keyword wildcard_folder_path: Microsoft Fabric Lakehouse Files wildcardFolderPath. Type: string (or Expression with resultType string). :paramtype wildcard_folder_path: JSON - :keyword wildcard_file_name: Microsoft Fabric LakeHouse Files wildcardFileName. Type: string + :keyword wildcard_file_name: Microsoft Fabric Lakehouse Files wildcardFileName. Type: string (or Expression with resultType string). :paramtype wildcard_file_name: JSON :keyword file_list_path: Point to a text file that lists each file (relative path to the path @@ -41086,7 +41247,7 @@ def __init__( class LakeHouseTableDataset(Dataset): - """Microsoft Fabric LakeHouse Table. + """Microsoft Fabric Lakehouse Table. All required parameters must be populated in order to send to server. @@ -41112,10 +41273,10 @@ class LakeHouseTableDataset(Dataset): :ivar folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. :vartype folder: ~azure.mgmt.datafactory.models.DatasetFolder - :ivar schema_type_properties_schema: The schema name of Microsoft Fabric LakeHouse Table. Type: + :ivar schema_type_properties_schema: The schema name of Microsoft Fabric Lakehouse Table. Type: string (or Expression with resultType string). :vartype schema_type_properties_schema: JSON - :ivar table: The name of Microsoft Fabric LakeHouse Table. Type: string (or Expression with + :ivar table: The name of Microsoft Fabric Lakehouse Table. Type: string (or Expression with resultType string). :vartype table: JSON """ @@ -41175,10 +41336,10 @@ def __init__( :keyword folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. :paramtype folder: ~azure.mgmt.datafactory.models.DatasetFolder - :keyword schema_type_properties_schema: The schema name of Microsoft Fabric LakeHouse Table. + :keyword schema_type_properties_schema: The schema name of Microsoft Fabric Lakehouse Table. Type: string (or Expression with resultType string). :paramtype schema_type_properties_schema: JSON - :keyword table: The name of Microsoft Fabric LakeHouse Table. Type: string (or Expression with + :keyword table: The name of Microsoft Fabric Lakehouse Table. Type: string (or Expression with resultType string). :paramtype table: JSON """ @@ -41193,13 +41354,13 @@ def __init__( folder=folder, **kwargs ) - self.type: str = "LakeHouseTable" + self.type: str = "LakehouseTable" self.schema_type_properties_schema = schema_type_properties_schema self.table = table class LakeHouseTableSink(CopySink): - """A copy activity for Microsoft Fabric LakeHouse Table sink. + """A copy activity for Microsoft Fabric Lakehouse Table sink. All required parameters must be populated in order to send to server. @@ -41226,7 +41387,7 @@ class LakeHouseTableSink(CopySink): :ivar disable_metrics_collection: If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). :vartype disable_metrics_collection: JSON - :ivar table_action_option: The type of table action for LakeHouse Table sink. Possible values + :ivar table_action_option: The type of table action for Lakehouse Table sink. Possible values include: "None", "Append", "Overwrite". :vartype table_action_option: JSON :ivar partition_option: Create partitions in folder structure based on one or multiple columns. @@ -41293,7 +41454,7 @@ def __init__( :keyword disable_metrics_collection: If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). :paramtype disable_metrics_collection: JSON - :keyword table_action_option: The type of table action for LakeHouse Table sink. Possible + :keyword table_action_option: The type of table action for Lakehouse Table sink. Possible values include: "None", "Append", "Overwrite". :paramtype table_action_option: JSON :keyword partition_option: Create partitions in folder structure based on one or multiple @@ -41321,7 +41482,7 @@ def __init__( class LakeHouseTableSource(CopySource): - """A copy activity source for Microsoft Fabric LakeHouse Table. + """A copy activity source for Microsoft Fabric Lakehouse Table. All required parameters must be populated in order to send to server. @@ -41423,7 +41584,7 @@ def __init__( class LakeHouseWriteSettings(StoreWriteSettings): - """Microsoft Fabric LakeHouse Files write settings. + """Microsoft Fabric Lakehouse Files write settings. All required parameters must be populated in order to send to server. @@ -48870,7 +49031,9 @@ def __init__( class OracleLinkedService(LinkedService): - """Oracle database. + """Oracle database. This linked service has supported version property. The Version 1.0 is + scheduled for deprecation while your pipeline will continue to run after EOL but without any + bug fix or new features. All required parameters must be populated in order to send to server. @@ -48890,10 +49053,59 @@ class OracleLinkedService(LinkedService): :ivar annotations: List of tags that can be used for describing the linked service. :vartype annotations: list[JSON] :ivar connection_string: The connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. Required. + AzureKeyVaultSecretReference. Only used for Version 1.0. :vartype connection_string: JSON + :ivar server: The location of Oracle database you want to connect to, the supported forms + include connector descriptor, Easy Connect (Plus) Naming and Oracle Net Services Name (Only + self-hosted IR). Type: string. Only used for Version 2.0. + :vartype server: JSON + :ivar authentication_type: Authentication type for connecting to the Oracle database. Only used + for Version 2.0. "Basic" + :vartype authentication_type: str or ~azure.mgmt.datafactory.models.OracleAuthenticationType + :ivar username: The Oracle database username. Type: string. Only used for Version 2.0. + :vartype username: JSON :ivar password: The Azure key vault secret reference of password in connection string. :vartype password: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference + :ivar encryption_client: Specifies the encryption client behavior. Supported values are + accepted, rejected, requested or required, default value is required. Type: string. Only used + for Version 2.0. + :vartype encryption_client: JSON + :ivar encryption_types_client: Specifies the encryption algorithms that client can use. + Supported values are AES128, AES192, AES256, 3DES112, 3DES168, default value is (AES256). Type: + string. Only used for Version 2.0. + :vartype encryption_types_client: JSON + :ivar crypto_checksum_client: Specifies the desired data integrity behavior when this client + connects to a server. Supported values are accepted, rejected, requested or required, default + value is required. Type: string. Only used for Version 2.0. + :vartype crypto_checksum_client: JSON + :ivar crypto_checksum_types_client: Specifies the crypto-checksum algorithms that client can + use. Supported values are SHA1, SHA256, SHA384, SHA512, default value is (SHA512). Type: + string. Only used for Version 2.0. + :vartype crypto_checksum_types_client: JSON + :ivar initial_lob_fetch_size: Specifies the amount that the source initially fetches for LOB + columns, default value is 0. Type: integer. Only used for Version 2.0. + :vartype initial_lob_fetch_size: JSON + :ivar fetch_size: Specifies the number of bytes that the driver allocates to fetch the data in + one database round-trip, default value is 10485760. Type: integer. Only used for Version 2.0. + :vartype fetch_size: JSON + :ivar statement_cache_size: Specifies the number of cursors or statements to be cached for each + database connection, default value is 0. Type: integer. Only used for Version 2.0. + :vartype statement_cache_size: JSON + :ivar initialization_string: Specifies a command that is issued immediately after connecting to + the database to manage session settings. Type: string. Only used for Version 2.0. + :vartype initialization_string: JSON + :ivar enable_bulk_load: Specifies whether to use bulk copy or batch insert when loading data + into the database, default value is true. Type: boolean. Only used for Version 2.0. + :vartype enable_bulk_load: JSON + :ivar support_v1_data_types: Specifies whether to use the Version 1.0 data type mappings. Do + not set this to true unless you want to keep backward compatibility with Version 1.0's data + type mappings, default value is false. Type: boolean. Only used for Version 2.0. + :vartype support_v1_data_types: JSON + :ivar fetch_tswtz_as_timestamp: Specifies whether the driver returns column value with the + TIMESTAMP WITH TIME ZONE data type as DateTime or string. This setting is ignored if + supportV1DataTypes is not true, default value is true. Type: boolean. Only used for Version + 2.0. + :vartype fetch_tswtz_as_timestamp: JSON :ivar encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string. :vartype encrypted_credential: str @@ -48901,7 +49113,6 @@ class OracleLinkedService(LinkedService): _validation = { "type": {"required": True}, - "connection_string": {"required": True}, } _attribute_map = { @@ -48913,21 +49124,49 @@ class OracleLinkedService(LinkedService): "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, "annotations": {"key": "annotations", "type": "[object]"}, "connection_string": {"key": "typeProperties.connectionString", "type": "object"}, + "server": {"key": "typeProperties.server", "type": "object"}, + "authentication_type": {"key": "typeProperties.authenticationType", "type": "str"}, + "username": {"key": "typeProperties.username", "type": "object"}, "password": {"key": "typeProperties.password", "type": "AzureKeyVaultSecretReference"}, + "encryption_client": {"key": "typeProperties.encryptionClient", "type": "object"}, + "encryption_types_client": {"key": "typeProperties.encryptionTypesClient", "type": "object"}, + "crypto_checksum_client": {"key": "typeProperties.cryptoChecksumClient", "type": "object"}, + "crypto_checksum_types_client": {"key": "typeProperties.cryptoChecksumTypesClient", "type": "object"}, + "initial_lob_fetch_size": {"key": "typeProperties.initialLobFetchSize", "type": "object"}, + "fetch_size": {"key": "typeProperties.fetchSize", "type": "object"}, + "statement_cache_size": {"key": "typeProperties.statementCacheSize", "type": "object"}, + "initialization_string": {"key": "typeProperties.initializationString", "type": "object"}, + "enable_bulk_load": {"key": "typeProperties.enableBulkLoad", "type": "object"}, + "support_v1_data_types": {"key": "typeProperties.supportV1DataTypes", "type": "object"}, + "fetch_tswtz_as_timestamp": {"key": "typeProperties.fetchTswtzAsTimestamp", "type": "object"}, "encrypted_credential": {"key": "typeProperties.encryptedCredential", "type": "str"}, } - def __init__( + def __init__( # pylint: disable=too-many-locals self, *, - connection_string: JSON, additional_properties: Optional[Dict[str, JSON]] = None, version: Optional[str] = None, connect_via: Optional["_models.IntegrationRuntimeReference"] = None, description: Optional[str] = None, parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, annotations: Optional[List[JSON]] = None, + connection_string: Optional[JSON] = None, + server: Optional[JSON] = None, + authentication_type: Optional[Union[str, "_models.OracleAuthenticationType"]] = None, + username: Optional[JSON] = None, password: Optional["_models.AzureKeyVaultSecretReference"] = None, + encryption_client: Optional[JSON] = None, + encryption_types_client: Optional[JSON] = None, + crypto_checksum_client: Optional[JSON] = None, + crypto_checksum_types_client: Optional[JSON] = None, + initial_lob_fetch_size: Optional[JSON] = None, + fetch_size: Optional[JSON] = None, + statement_cache_size: Optional[JSON] = None, + initialization_string: Optional[JSON] = None, + enable_bulk_load: Optional[JSON] = None, + support_v1_data_types: Optional[JSON] = None, + fetch_tswtz_as_timestamp: Optional[JSON] = None, encrypted_credential: Optional[str] = None, **kwargs: Any ) -> None: @@ -48946,10 +49185,60 @@ def __init__( :keyword annotations: List of tags that can be used for describing the linked service. :paramtype annotations: list[JSON] :keyword connection_string: The connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. Required. + AzureKeyVaultSecretReference. Only used for Version 1.0. :paramtype connection_string: JSON + :keyword server: The location of Oracle database you want to connect to, the supported forms + include connector descriptor, Easy Connect (Plus) Naming and Oracle Net Services Name (Only + self-hosted IR). Type: string. Only used for Version 2.0. + :paramtype server: JSON + :keyword authentication_type: Authentication type for connecting to the Oracle database. Only + used for Version 2.0. "Basic" + :paramtype authentication_type: str or ~azure.mgmt.datafactory.models.OracleAuthenticationType + :keyword username: The Oracle database username. Type: string. Only used for Version 2.0. + :paramtype username: JSON :keyword password: The Azure key vault secret reference of password in connection string. :paramtype password: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference + :keyword encryption_client: Specifies the encryption client behavior. Supported values are + accepted, rejected, requested or required, default value is required. Type: string. Only used + for Version 2.0. + :paramtype encryption_client: JSON + :keyword encryption_types_client: Specifies the encryption algorithms that client can use. + Supported values are AES128, AES192, AES256, 3DES112, 3DES168, default value is (AES256). Type: + string. Only used for Version 2.0. + :paramtype encryption_types_client: JSON + :keyword crypto_checksum_client: Specifies the desired data integrity behavior when this client + connects to a server. Supported values are accepted, rejected, requested or required, default + value is required. Type: string. Only used for Version 2.0. + :paramtype crypto_checksum_client: JSON + :keyword crypto_checksum_types_client: Specifies the crypto-checksum algorithms that client can + use. Supported values are SHA1, SHA256, SHA384, SHA512, default value is (SHA512). Type: + string. Only used for Version 2.0. + :paramtype crypto_checksum_types_client: JSON + :keyword initial_lob_fetch_size: Specifies the amount that the source initially fetches for LOB + columns, default value is 0. Type: integer. Only used for Version 2.0. + :paramtype initial_lob_fetch_size: JSON + :keyword fetch_size: Specifies the number of bytes that the driver allocates to fetch the data + in one database round-trip, default value is 10485760. Type: integer. Only used for Version + 2.0. + :paramtype fetch_size: JSON + :keyword statement_cache_size: Specifies the number of cursors or statements to be cached for + each database connection, default value is 0. Type: integer. Only used for Version 2.0. + :paramtype statement_cache_size: JSON + :keyword initialization_string: Specifies a command that is issued immediately after connecting + to the database to manage session settings. Type: string. Only used for Version 2.0. + :paramtype initialization_string: JSON + :keyword enable_bulk_load: Specifies whether to use bulk copy or batch insert when loading data + into the database, default value is true. Type: boolean. Only used for Version 2.0. + :paramtype enable_bulk_load: JSON + :keyword support_v1_data_types: Specifies whether to use the Version 1.0 data type mappings. Do + not set this to true unless you want to keep backward compatibility with Version 1.0's data + type mappings, default value is false. Type: boolean. Only used for Version 2.0. + :paramtype support_v1_data_types: JSON + :keyword fetch_tswtz_as_timestamp: Specifies whether the driver returns column value with the + TIMESTAMP WITH TIME ZONE data type as DateTime or string. This setting is ignored if + supportV1DataTypes is not true, default value is true. Type: boolean. Only used for Version + 2.0. + :paramtype fetch_tswtz_as_timestamp: JSON :keyword encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string. :paramtype encrypted_credential: str @@ -48965,7 +49254,21 @@ def __init__( ) self.type: str = "Oracle" self.connection_string = connection_string + self.server = server + self.authentication_type = authentication_type + self.username = username self.password = password + self.encryption_client = encryption_client + self.encryption_types_client = encryption_types_client + self.crypto_checksum_client = crypto_checksum_client + self.crypto_checksum_types_client = crypto_checksum_types_client + self.initial_lob_fetch_size = initial_lob_fetch_size + self.fetch_size = fetch_size + self.statement_cache_size = statement_cache_size + self.initialization_string = initialization_string + self.enable_bulk_load = enable_bulk_load + self.support_v1_data_types = support_v1_data_types + self.fetch_tswtz_as_timestamp = fetch_tswtz_as_timestamp self.encrypted_credential = encrypted_credential @@ -50084,9 +50387,9 @@ class OrcWriteSettings(FormatWriteSettings): :ivar max_rows_per_file: Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). :vartype max_rows_per_file: JSON - :ivar file_name_prefix: Specifies the file name pattern - :code:``_:code:``.:code:`` when copy from non-file - based store without partitionOptions. Type: string (or Expression with resultType string). + :ivar file_name_prefix: Specifies the file name pattern :code:``_\\ + :code:``.\\ :code:`` when copy from non-file based store without + partitionOptions. Type: string (or Expression with resultType string). :vartype file_name_prefix: JSON """ @@ -50116,9 +50419,9 @@ def __init__( :keyword max_rows_per_file: Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). :paramtype max_rows_per_file: JSON - :keyword file_name_prefix: Specifies the file name pattern - :code:``_:code:``.:code:`` when copy from non-file - based store without partitionOptions. Type: string (or Expression with resultType string). + :keyword file_name_prefix: Specifies the file name pattern :code:``_\\ + :code:``.\\ :code:`` when copy from non-file based store without + partitionOptions. Type: string (or Expression with resultType string). :paramtype file_name_prefix: JSON """ super().__init__(additional_properties=additional_properties, **kwargs) @@ -50638,9 +50941,9 @@ class ParquetWriteSettings(FormatWriteSettings): :ivar max_rows_per_file: Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). :vartype max_rows_per_file: JSON - :ivar file_name_prefix: Specifies the file name pattern - :code:``_:code:``.:code:`` when copy from non-file - based store without partitionOptions. Type: string (or Expression with resultType string). + :ivar file_name_prefix: Specifies the file name pattern :code:``_\\ + :code:``.\\ :code:`` when copy from non-file based store without + partitionOptions. Type: string (or Expression with resultType string). :vartype file_name_prefix: JSON """ @@ -50670,9 +50973,9 @@ def __init__( :keyword max_rows_per_file: Limit the written file's row count to be smaller than or equal to the specified count. Type: integer (or Expression with resultType integer). :paramtype max_rows_per_file: JSON - :keyword file_name_prefix: Specifies the file name pattern - :code:``_:code:``.:code:`` when copy from non-file - based store without partitionOptions. Type: string (or Expression with resultType string). + :keyword file_name_prefix: Specifies the file name pattern :code:``_\\ + :code:``.\\ :code:`` when copy from non-file based store without + partitionOptions. Type: string (or Expression with resultType string). :paramtype file_name_prefix: JSON """ super().__init__(additional_properties=additional_properties, **kwargs) @@ -59807,8 +60110,8 @@ class SapOdpLinkedService(LinkedService): :ivar message_server_service: The service name or port number of the Message Server. Type: string (or Expression with resultType string). :vartype message_server_service: JSON - :ivar snc_mode: SNC activation indicator to access the SAP server where the table is located. - Must be either 0 (off) or 1 (on). Type: string (or Expression with resultType string). + :ivar snc_mode: SNC activation flag (Boolean) to access the SAP server where the table is + located. Type: boolean (or Expression with resultType boolean). :vartype snc_mode: JSON :ivar snc_my_name: Initiator's SNC name to access the SAP server where the table is located. Type: string (or Expression with resultType string). @@ -59939,8 +60242,8 @@ def __init__( # pylint: disable=too-many-locals :keyword message_server_service: The service name or port number of the Message Server. Type: string (or Expression with resultType string). :paramtype message_server_service: JSON - :keyword snc_mode: SNC activation indicator to access the SAP server where the table is - located. Must be either 0 (off) or 1 (on). Type: string (or Expression with resultType string). + :keyword snc_mode: SNC activation flag (Boolean) to access the SAP server where the table is + located. Type: boolean (or Expression with resultType boolean). :paramtype snc_mode: JSON :keyword snc_my_name: Initiator's SNC name to access the SAP server where the table is located. Type: string (or Expression with resultType string). @@ -60723,8 +61026,8 @@ class SapTableLinkedService(LinkedService): :ivar message_server_service: The service name or port number of the Message Server. Type: string (or Expression with resultType string). :vartype message_server_service: JSON - :ivar snc_mode: SNC activation indicator to access the SAP server where the table is located. - Must be either 0 (off) or 1 (on). Type: string (or Expression with resultType string). + :ivar snc_mode: SNC activation flag (Boolean) to access the SAP server where the table is + located. Type: boolean (or Expression with resultType boolean). :vartype snc_mode: JSON :ivar snc_my_name: Initiator's SNC name to access the SAP server where the table is located. Type: string (or Expression with resultType string). @@ -60845,8 +61148,8 @@ def __init__( :keyword message_server_service: The service name or port number of the Message Server. Type: string (or Expression with resultType string). :paramtype message_server_service: JSON - :keyword snc_mode: SNC activation indicator to access the SAP server where the table is - located. Must be either 0 (off) or 1 (on). Type: string (or Expression with resultType string). + :keyword snc_mode: SNC activation flag (Boolean) to access the SAP server where the table is + located. Type: boolean (or Expression with resultType boolean). :paramtype snc_mode: JSON :keyword snc_my_name: Initiator's SNC name to access the SAP server where the table is located. Type: string (or Expression with resultType string). @@ -61465,6 +61768,10 @@ class ScriptActivity(ExecutionActivity): :vartype scripts: list[~azure.mgmt.datafactory.models.ScriptActivityScriptBlock] :ivar log_settings: Log settings of script activity. :vartype log_settings: ~azure.mgmt.datafactory.models.ScriptActivityTypePropertiesLogSettings + :ivar return_multistatement_result: Enable to retrieve result sets from multiple SQL statements + and the number of rows affected by the DML statement. Supported connector: SnowflakeV2. Type: + boolean (or Expression with resultType boolean). + :vartype return_multistatement_result: JSON """ _validation = { @@ -61486,6 +61793,7 @@ class ScriptActivity(ExecutionActivity): "script_block_execution_timeout": {"key": "typeProperties.scriptBlockExecutionTimeout", "type": "object"}, "scripts": {"key": "typeProperties.scripts", "type": "[ScriptActivityScriptBlock]"}, "log_settings": {"key": "typeProperties.logSettings", "type": "ScriptActivityTypePropertiesLogSettings"}, + "return_multistatement_result": {"key": "typeProperties.returnMultistatementResult", "type": "object"}, } def __init__( @@ -61503,6 +61811,7 @@ def __init__( script_block_execution_timeout: Optional[JSON] = None, scripts: Optional[List["_models.ScriptActivityScriptBlock"]] = None, log_settings: Optional["_models.ScriptActivityTypePropertiesLogSettings"] = None, + return_multistatement_result: Optional[JSON] = None, **kwargs: Any ) -> None: """ @@ -61536,6 +61845,10 @@ def __init__( :paramtype scripts: list[~azure.mgmt.datafactory.models.ScriptActivityScriptBlock] :keyword log_settings: Log settings of script activity. :paramtype log_settings: ~azure.mgmt.datafactory.models.ScriptActivityTypePropertiesLogSettings + :keyword return_multistatement_result: Enable to retrieve result sets from multiple SQL + statements and the number of rows affected by the DML statement. Supported connector: + SnowflakeV2. Type: boolean (or Expression with resultType boolean). + :paramtype return_multistatement_result: JSON """ super().__init__( additional_properties=additional_properties, @@ -61553,6 +61866,7 @@ def __init__( self.script_block_execution_timeout = script_block_execution_timeout self.scripts = scripts self.log_settings = log_settings + self.return_multistatement_result = return_multistatement_result class ScriptActivityParameter(_serialization.Model): @@ -71048,6 +71362,55 @@ def __init__( self.encrypted_credential = encrypted_credential +class TeradataImportCommand(ImportSettings): + """Teradata import command settings. + + All required parameters must be populated in order to send to server. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: The import setting type. Required. + :vartype type: str + :ivar additional_format_options: Additional format options for Teradata Copy Command. The + format options only applies to direct copy from CSV source. Type: key value pairs (value should + be string type) (or Expression with resultType object). Example: "additionalFormatOptions": { + "timeFormat": "HHhMImSSs" }. + :vartype additional_format_options: JSON + """ + + _validation = { + "type": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "additional_format_options": {"key": "additionalFormatOptions", "type": "object"}, + } + + def __init__( + self, + *, + additional_properties: Optional[Dict[str, JSON]] = None, + additional_format_options: Optional[JSON] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword additional_format_options: Additional format options for Teradata Copy Command. The + format options only applies to direct copy from CSV source. Type: key value pairs (value should + be string type) (or Expression with resultType object). Example: "additionalFormatOptions": { + "timeFormat": "HHhMImSSs" }. + :paramtype additional_format_options: JSON + """ + super().__init__(additional_properties=additional_properties, **kwargs) + self.type: str = "TeradataImportCommand" + self.additional_format_options = additional_format_options + + class TeradataLinkedService(LinkedService): """Linked service for Teradata data source. @@ -71069,7 +71432,7 @@ class TeradataLinkedService(LinkedService): :ivar annotations: List of tags that can be used for describing the linked service. :vartype annotations: list[JSON] :ivar connection_string: Teradata ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. + AzureKeyVaultSecretReference. Only applied for version 1.0. :vartype connection_string: JSON :ivar server: Server name for connection. Type: string (or Expression with resultType string). :vartype server: JSON @@ -71081,6 +71444,28 @@ class TeradataLinkedService(LinkedService): :vartype username: JSON :ivar password: Password for authentication. :vartype password: ~azure.mgmt.datafactory.models.SecretBase + :ivar ssl_mode: SSL mode for connection. Valid values including: “Disable”, “Allow”, “Prefer”, + “Require”, “Verify-CA”, “Verify-Full”. Default value is “Verify-Full”. Type: string (or + Expression with resultType string). Only applied for version 2.0. + :vartype ssl_mode: JSON + :ivar port_number: The port numbers when connecting to server through non HTTPS/TLS + connections. Type: integer (or Expression with resultType integer). Only used for V2. Only + applied for version 2.0. + :vartype port_number: JSON + :ivar https_port_number: The port numbers when connecting to server through HTTPS/TLS + connections. Type: integer (or Expression with resultType integer). Only applied for version + 2.0. + :vartype https_port_number: JSON + :ivar use_data_encryption: Specifies whether to encrypt all communication with the Teradata + database. Allowed values are 0 or 1. This setting will be ignored for HTTPS/TLS connections. + Type: integer (or Expression with resultType integer). Only applied for version 2.0. + :vartype use_data_encryption: JSON + :ivar character_set: The character set to use for the connection. Type: string (or Expression + with resultType string). Only applied for version 2.0. + :vartype character_set: JSON + :ivar max_resp_size: The maximum size of the response buffer for SQL requests, in bytes. Type: + integer. Only applied for version 2.0. + :vartype max_resp_size: JSON :ivar encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string. :vartype encrypted_credential: str @@ -71103,6 +71488,12 @@ class TeradataLinkedService(LinkedService): "authentication_type": {"key": "typeProperties.authenticationType", "type": "str"}, "username": {"key": "typeProperties.username", "type": "object"}, "password": {"key": "typeProperties.password", "type": "SecretBase"}, + "ssl_mode": {"key": "typeProperties.sslMode", "type": "object"}, + "port_number": {"key": "typeProperties.portNumber", "type": "object"}, + "https_port_number": {"key": "typeProperties.httpsPortNumber", "type": "object"}, + "use_data_encryption": {"key": "typeProperties.useDataEncryption", "type": "object"}, + "character_set": {"key": "typeProperties.characterSet", "type": "object"}, + "max_resp_size": {"key": "typeProperties.maxRespSize", "type": "object"}, "encrypted_credential": {"key": "typeProperties.encryptedCredential", "type": "str"}, } @@ -71120,6 +71511,12 @@ def __init__( authentication_type: Optional[Union[str, "_models.TeradataAuthenticationType"]] = None, username: Optional[JSON] = None, password: Optional["_models.SecretBase"] = None, + ssl_mode: Optional[JSON] = None, + port_number: Optional[JSON] = None, + https_port_number: Optional[JSON] = None, + use_data_encryption: Optional[JSON] = None, + character_set: Optional[JSON] = None, + max_resp_size: Optional[JSON] = None, encrypted_credential: Optional[str] = None, **kwargs: Any ) -> None: @@ -71138,7 +71535,7 @@ def __init__( :keyword annotations: List of tags that can be used for describing the linked service. :paramtype annotations: list[JSON] :keyword connection_string: Teradata ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. + AzureKeyVaultSecretReference. Only applied for version 1.0. :paramtype connection_string: JSON :keyword server: Server name for connection. Type: string (or Expression with resultType string). @@ -71152,6 +71549,28 @@ def __init__( :paramtype username: JSON :keyword password: Password for authentication. :paramtype password: ~azure.mgmt.datafactory.models.SecretBase + :keyword ssl_mode: SSL mode for connection. Valid values including: “Disable”, “Allow”, + “Prefer”, “Require”, “Verify-CA”, “Verify-Full”. Default value is “Verify-Full”. Type: string + (or Expression with resultType string). Only applied for version 2.0. + :paramtype ssl_mode: JSON + :keyword port_number: The port numbers when connecting to server through non HTTPS/TLS + connections. Type: integer (or Expression with resultType integer). Only used for V2. Only + applied for version 2.0. + :paramtype port_number: JSON + :keyword https_port_number: The port numbers when connecting to server through HTTPS/TLS + connections. Type: integer (or Expression with resultType integer). Only applied for version + 2.0. + :paramtype https_port_number: JSON + :keyword use_data_encryption: Specifies whether to encrypt all communication with the Teradata + database. Allowed values are 0 or 1. This setting will be ignored for HTTPS/TLS connections. + Type: integer (or Expression with resultType integer). Only applied for version 2.0. + :paramtype use_data_encryption: JSON + :keyword character_set: The character set to use for the connection. Type: string (or + Expression with resultType string). Only applied for version 2.0. + :paramtype character_set: JSON + :keyword max_resp_size: The maximum size of the response buffer for SQL requests, in bytes. + Type: integer. Only applied for version 2.0. + :paramtype max_resp_size: JSON :keyword encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string. :paramtype encrypted_credential: str @@ -71171,6 +71590,12 @@ def __init__( self.authentication_type = authentication_type self.username = username self.password = password + self.ssl_mode = ssl_mode + self.port_number = port_number + self.https_port_number = https_port_number + self.use_data_encryption = use_data_encryption + self.character_set = character_set + self.max_resp_size = max_resp_size self.encrypted_credential = encrypted_credential @@ -71223,6 +71648,106 @@ def __init__( self.partition_lower_bound = partition_lower_bound +class TeradataSink(CopySink): + """A copy activity Teradata sink. + + All required parameters must be populated in order to send to server. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: Copy sink type. Required. + :vartype type: str + :ivar write_batch_size: Write batch size. Type: integer (or Expression with resultType + integer), minimum: 0. + :vartype write_batch_size: JSON + :ivar write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType + string), pattern: ((\\d+).)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype write_batch_timeout: JSON + :ivar sink_retry_count: Sink retry count. Type: integer (or Expression with resultType + integer). + :vartype sink_retry_count: JSON + :ivar sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), + pattern: ((\\d+).)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype sink_retry_wait: JSON + :ivar max_concurrent_connections: The maximum concurrent connection count for the sink data + store. Type: integer (or Expression with resultType integer). + :vartype max_concurrent_connections: JSON + :ivar disable_metrics_collection: If true, disable data store metrics collection. Default is + false. Type: boolean (or Expression with resultType boolean). + :vartype disable_metrics_collection: JSON + :ivar import_settings: Teradata import settings. + :vartype import_settings: ~azure.mgmt.datafactory.models.TeradataImportCommand + """ + + _validation = { + "type": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "write_batch_size": {"key": "writeBatchSize", "type": "object"}, + "write_batch_timeout": {"key": "writeBatchTimeout", "type": "object"}, + "sink_retry_count": {"key": "sinkRetryCount", "type": "object"}, + "sink_retry_wait": {"key": "sinkRetryWait", "type": "object"}, + "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, + "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, + "import_settings": {"key": "importSettings", "type": "TeradataImportCommand"}, + } + + def __init__( + self, + *, + additional_properties: Optional[Dict[str, JSON]] = None, + write_batch_size: Optional[JSON] = None, + write_batch_timeout: Optional[JSON] = None, + sink_retry_count: Optional[JSON] = None, + sink_retry_wait: Optional[JSON] = None, + max_concurrent_connections: Optional[JSON] = None, + disable_metrics_collection: Optional[JSON] = None, + import_settings: Optional["_models.TeradataImportCommand"] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword write_batch_size: Write batch size. Type: integer (or Expression with resultType + integer), minimum: 0. + :paramtype write_batch_size: JSON + :keyword write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType + string), pattern: ((\\d+).)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype write_batch_timeout: JSON + :keyword sink_retry_count: Sink retry count. Type: integer (or Expression with resultType + integer). + :paramtype sink_retry_count: JSON + :keyword sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), + pattern: ((\\d+).)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype sink_retry_wait: JSON + :keyword max_concurrent_connections: The maximum concurrent connection count for the sink data + store. Type: integer (or Expression with resultType integer). + :paramtype max_concurrent_connections: JSON + :keyword disable_metrics_collection: If true, disable data store metrics collection. Default is + false. Type: boolean (or Expression with resultType boolean). + :paramtype disable_metrics_collection: JSON + :keyword import_settings: Teradata import settings. + :paramtype import_settings: ~azure.mgmt.datafactory.models.TeradataImportCommand + """ + super().__init__( + additional_properties=additional_properties, + write_batch_size=write_batch_size, + write_batch_timeout=write_batch_timeout, + sink_retry_count=sink_retry_count, + sink_retry_wait=sink_retry_wait, + max_concurrent_connections=max_concurrent_connections, + disable_metrics_collection=disable_metrics_collection, + **kwargs + ) + self.type: str = "TeradataSink" + self.import_settings = import_settings + + class TeradataSource(TabularSource): """A copy activity Teradata source. @@ -72288,6 +72813,12 @@ class TypeConversionSettings(_serialization.Model): :ivar time_span_format: The format for TimeSpan values. Type: string (or Expression with resultType string). :vartype time_span_format: JSON + :ivar time_format: The format for Time values. Type: string (or Expression with resultType + string). + :vartype time_format: JSON + :ivar date_format: The format for Date values. Type: string (or Expression with resultType + string). + :vartype date_format: JSON :ivar culture: The culture used to convert data from/to string. Type: string (or Expression with resultType string). :vartype culture: JSON @@ -72299,6 +72830,8 @@ class TypeConversionSettings(_serialization.Model): "date_time_format": {"key": "dateTimeFormat", "type": "object"}, "date_time_offset_format": {"key": "dateTimeOffsetFormat", "type": "object"}, "time_span_format": {"key": "timeSpanFormat", "type": "object"}, + "time_format": {"key": "timeFormat", "type": "object"}, + "date_format": {"key": "dateFormat", "type": "object"}, "culture": {"key": "culture", "type": "object"}, } @@ -72310,6 +72843,8 @@ def __init__( date_time_format: Optional[JSON] = None, date_time_offset_format: Optional[JSON] = None, time_span_format: Optional[JSON] = None, + time_format: Optional[JSON] = None, + date_format: Optional[JSON] = None, culture: Optional[JSON] = None, **kwargs: Any ) -> None: @@ -72329,6 +72864,12 @@ def __init__( :keyword time_span_format: The format for TimeSpan values. Type: string (or Expression with resultType string). :paramtype time_span_format: JSON + :keyword time_format: The format for Time values. Type: string (or Expression with resultType + string). + :paramtype time_format: JSON + :keyword date_format: The format for Date values. Type: string (or Expression with resultType + string). + :paramtype date_format: JSON :keyword culture: The culture used to convert data from/to string. Type: string (or Expression with resultType string). :paramtype culture: JSON @@ -72339,6 +72880,8 @@ def __init__( self.date_time_format = date_time_format self.date_time_offset_format = date_time_offset_format self.time_span_format = time_span_format + self.time_format = time_format + self.date_format = date_format self.culture = culture diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/_factories_operations.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/_factories_operations.py index 6656f49415bd..b665fb019a50 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/_factories_operations.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/_factories_operations.py @@ -40,30 +40,6 @@ _SERIALIZER.client_side_validation = False -def build_list_request(subscription_id: str, **kwargs: Any) -> HttpRequest: - _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) - _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2018-06-01")) - accept = _headers.pop("Accept", "application/json") - - # Construct URL - _url = kwargs.pop("template_url", "/subscriptions/{subscriptionId}/providers/Microsoft.DataFactory/factories") - path_format_arguments = { - "subscriptionId": _SERIALIZER.url("subscription_id", subscription_id, "str"), - } - - _url: str = _url.format(**path_format_arguments) # type: ignore - - # Construct parameters - _params["api-version"] = _SERIALIZER.query("api_version", api_version, "str") - - # Construct headers - _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") - - return HttpRequest(method="GET", url=_url, params=_params, headers=_headers, **kwargs) - - def build_configure_factory_repo_request(location_id: str, subscription_id: str, **kwargs: Any) -> HttpRequest: _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) @@ -405,80 +381,6 @@ def __init__(self, *args, **kwargs): self._serialize = input_args.pop(0) if input_args else kwargs.pop("serializer") self._deserialize = input_args.pop(0) if input_args else kwargs.pop("deserializer") - @distributed_trace - def list(self, **kwargs: Any) -> Iterable["_models.Factory"]: - """Lists factories under the specified subscription. - - :return: An iterator like instance of either Factory or the result of cls(response) - :rtype: ~azure.core.paging.ItemPaged[~azure.mgmt.datafactory.models.Factory] - :raises ~azure.core.exceptions.HttpResponseError: - """ - _headers = kwargs.pop("headers", {}) or {} - _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - - api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._config.api_version)) - cls: ClsType[_models.FactoryListResponse] = kwargs.pop("cls", None) - - error_map: MutableMapping = { - 401: ClientAuthenticationError, - 404: ResourceNotFoundError, - 409: ResourceExistsError, - 304: ResourceNotModifiedError, - } - error_map.update(kwargs.pop("error_map", {}) or {}) - - def prepare_request(next_link=None): - if not next_link: - - _request = build_list_request( - subscription_id=self._config.subscription_id, - api_version=api_version, - headers=_headers, - params=_params, - ) - _request.url = self._client.format_url(_request.url) - - else: - # make call to next link with the client's api-version - _parsed_next_link = urllib.parse.urlparse(next_link) - _next_request_params = case_insensitive_dict( - { - key: [urllib.parse.quote(v) for v in value] - for key, value in urllib.parse.parse_qs(_parsed_next_link.query).items() - } - ) - _next_request_params["api-version"] = self._config.api_version - _request = HttpRequest( - "GET", urllib.parse.urljoin(next_link, _parsed_next_link.path), params=_next_request_params - ) - _request.url = self._client.format_url(_request.url) - _request.method = "GET" - return _request - - def extract_data(pipeline_response): - deserialized = self._deserialize("FactoryListResponse", pipeline_response) - list_of_elem = deserialized.value - if cls: - list_of_elem = cls(list_of_elem) # type: ignore - return deserialized.next_link or None, iter(list_of_elem) - - def get_next(next_link=None): - _request = prepare_request(next_link) - - _stream = False - pipeline_response: PipelineResponse = self._client._pipeline.run( # pylint: disable=protected-access - _request, stream=_stream, **kwargs - ) - response = pipeline_response.http_response - - if response.status_code not in [200]: - map_error(status_code=response.status_code, response=response, error_map=error_map) - raise HttpResponseError(response=response, error_format=ARMErrorFormat) - - return pipeline_response - - return ItemPaged(get_next, extract_data) - @overload def configure_factory_repo( self, diff --git a/sdk/datafactory/azure-mgmt-datafactory/generated_samples/factories_list.py b/sdk/datafactory/azure-mgmt-datafactory/generated_samples/factories_list.py deleted file mode 100644 index eea9551b7ce0..000000000000 --- a/sdk/datafactory/azure-mgmt-datafactory/generated_samples/factories_list.py +++ /dev/null @@ -1,40 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for license information. -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is regenerated. -# -------------------------------------------------------------------------- - -from azure.identity import DefaultAzureCredential - -from azure.mgmt.datafactory import DataFactoryManagementClient - -""" -# PREREQUISITES - pip install azure-identity - pip install azure-mgmt-datafactory -# USAGE - python factories_list.py - - Before run the sample, please set the values of the client ID, tenant ID and client secret - of the AAD application as environment variables: AZURE_CLIENT_ID, AZURE_TENANT_ID, - AZURE_CLIENT_SECRET. For more info about how to get the value, please see: - https://docs.microsoft.com/azure/active-directory/develop/howto-create-service-principal-portal -""" - - -def main(): - client = DataFactoryManagementClient( - credential=DefaultAzureCredential(), - subscription_id="12345678-1234-1234-1234-12345678abc", - ) - - response = client.factories.list() - for item in response: - print(item) - - -# x-ms-original-file: specification/datafactory/resource-manager/Microsoft.DataFactory/stable/2018-06-01/examples/Factories_List.json -if __name__ == "__main__": - main() diff --git a/sdk/datafactory/azure-mgmt-datafactory/generated_tests/test_data_factory_management_factories_operations.py b/sdk/datafactory/azure-mgmt-datafactory/generated_tests/test_data_factory_management_factories_operations.py index 27215817e5e6..b8a222908d4a 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/generated_tests/test_data_factory_management_factories_operations.py +++ b/sdk/datafactory/azure-mgmt-datafactory/generated_tests/test_data_factory_management_factories_operations.py @@ -18,16 +18,6 @@ class TestDataFactoryManagementFactoriesOperations(AzureMgmtRecordedTestCase): def setup_method(self, method): self.client = self.create_mgmt_client(DataFactoryManagementClient) - @RandomNameResourceGroupPreparer(location=AZURE_LOCATION) - @recorded_by_proxy - def test_factories_list(self, resource_group): - response = self.client.factories.list( - api_version="2018-06-01", - ) - result = [r for r in response] - # please add some check logic here by yourself - # ... - @RandomNameResourceGroupPreparer(location=AZURE_LOCATION) @recorded_by_proxy def test_factories_configure_factory_repo(self, resource_group): diff --git a/sdk/datafactory/azure-mgmt-datafactory/generated_tests/test_data_factory_management_factories_operations_async.py b/sdk/datafactory/azure-mgmt-datafactory/generated_tests/test_data_factory_management_factories_operations_async.py index 6a41ac09f6e9..7e0d25f603f7 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/generated_tests/test_data_factory_management_factories_operations_async.py +++ b/sdk/datafactory/azure-mgmt-datafactory/generated_tests/test_data_factory_management_factories_operations_async.py @@ -19,16 +19,6 @@ class TestDataFactoryManagementFactoriesOperationsAsync(AzureMgmtRecordedTestCas def setup_method(self, method): self.client = self.create_mgmt_client(DataFactoryManagementClient, is_async=True) - @RandomNameResourceGroupPreparer(location=AZURE_LOCATION) - @recorded_by_proxy_async - async def test_factories_list(self, resource_group): - response = self.client.factories.list( - api_version="2018-06-01", - ) - result = [r async for r in response] - # please add some check logic here by yourself - # ... - @RandomNameResourceGroupPreparer(location=AZURE_LOCATION) @recorded_by_proxy_async async def test_factories_configure_factory_repo(self, resource_group): diff --git a/sdk/datafactory/azure-mgmt-datafactory/setup.py b/sdk/datafactory/azure-mgmt-datafactory/setup.py index 570262824209..907fd55998d6 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/setup.py +++ b/sdk/datafactory/azure-mgmt-datafactory/setup.py @@ -22,11 +22,9 @@ # Version extraction inspired from 'requests' with open( - ( - os.path.join(package_folder_path, "version.py") - if os.path.exists(os.path.join(package_folder_path, "version.py")) - else os.path.join(package_folder_path, "_version.py") - ), + os.path.join(package_folder_path, "version.py") + if os.path.exists(os.path.join(package_folder_path, "version.py")) + else os.path.join(package_folder_path, "_version.py"), "r", ) as fd: version = re.search(r'^VERSION\s*=\s*[\'"]([^\'"]*)[\'"]', fd.read(), re.MULTILINE).group(1)