From a81a7854a39cc32839eb3a922270a4f89dd9a82b Mon Sep 17 00:00:00 2001 From: azure-sdk Date: Tue, 24 Feb 2026 08:49:17 +0000 Subject: [PATCH] code and test --- .../azure-mgmt-storagecache/CHANGELOG.md | 15 + .../azure-mgmt-storagecache/_metadata.json | 8 +- .../apiview-properties.json | 15 + .../azure/mgmt/storagecache/_configuration.py | 4 +- .../_storage_cache_management_client.py | 6 +- .../mgmt/storagecache/_utils/serialization.py | 19 +- .../azure/mgmt/storagecache/_version.py | 2 +- .../mgmt/storagecache/aio/_configuration.py | 4 +- .../aio/_storage_cache_management_client.py | 6 +- .../storagecache/aio/operations/__init__.py | 2 + .../operations/_aml_filesystems_operations.py | 2 +- .../operations/_asc_operations_operations.py | 1 + .../aio/operations/_asc_usages_operations.py | 2 +- .../_auto_export_jobs_operations.py | 21 +- .../_auto_import_jobs_operations.py | 21 +- .../aio/operations/_caches_operations.py | 2 +- .../operations/_expansion_jobs_operations.py | 797 +++++++++++++ .../aio/operations/_import_jobs_operations.py | 21 +- .../aio/operations/_operations.py | 2 +- .../aio/operations/_skus_operations.py | 2 +- ...rage_cache_management_client_operations.py | 6 +- .../operations/_storage_target_operations.py | 1 + .../operations/_storage_targets_operations.py | 1 + .../operations/_usage_models_operations.py | 2 +- .../mgmt/storagecache/models/__init__.py | 10 + .../mgmt/storagecache/models/_models_py3.py | 177 ++- .../_storage_cache_management_client_enums.py | 27 + .../mgmt/storagecache/operations/__init__.py | 2 + .../operations/_aml_filesystems_operations.py | 18 +- .../operations/_asc_operations_operations.py | 3 +- .../operations/_asc_usages_operations.py | 4 +- .../_auto_export_jobs_operations.py | 31 +- .../_auto_import_jobs_operations.py | 31 +- .../operations/_caches_operations.py | 34 +- .../operations/_expansion_jobs_operations.py | 1029 +++++++++++++++++ .../operations/_import_jobs_operations.py | 31 +- .../storagecache/operations/_operations.py | 4 +- .../operations/_skus_operations.py | 4 +- ...rage_cache_management_client_operations.py | 10 +- .../operations/_storage_target_operations.py | 9 +- .../operations/_storage_targets_operations.py | 13 +- .../operations/_usage_models_operations.py | 4 +- .../aml_filesystems_archive.py | 2 +- .../aml_filesystems_cancel_archive.py | 2 +- .../aml_filesystems_create_or_update.py | 2 +- .../aml_filesystems_delete.py | 2 +- .../generated_samples/aml_filesystems_get.py | 2 +- .../generated_samples/aml_filesystems_list.py | 2 +- .../aml_filesystems_list_by_resource_group.py | 2 +- .../aml_filesystems_update.py | 2 +- .../generated_samples/asc_operations_get.py | 2 +- .../asc_resource_usages_get.py | 2 +- .../auto_export_jobs_create_or_update.py | 2 +- .../auto_export_jobs_delete.py | 2 +- .../generated_samples/auto_export_jobs_get.py | 2 +- ...auto_export_jobs_list_by_aml_filesystem.py | 2 +- .../auto_export_jobs_update.py | 2 +- .../auto_import_jobs_create_or_update.py | 2 +- .../auto_import_jobs_delete.py | 2 +- .../generated_samples/auto_import_jobs_get.py | 2 +- ...auto_import_jobs_list_by_aml_filesystem.py | 2 +- .../auto_import_jobs_update.py | 2 +- .../caches_create_or_update.py | 2 +- .../caches_create_or_update_ldap_only.py | 2 +- .../generated_samples/caches_debug_info.py | 2 +- .../generated_samples/caches_delete.py | 2 +- .../generated_samples/caches_flush.py | 2 +- .../generated_samples/caches_get.py | 2 +- .../generated_samples/caches_list.py | 2 +- .../caches_list_by_resource_group.py | 2 +- .../generated_samples/caches_start.py | 2 +- .../generated_samples/caches_stop.py | 2 +- .../generated_samples/caches_update.py | 2 +- .../caches_update_ldap_only.py | 2 +- .../caches_upgrade_firmware.py | 2 +- .../generated_samples/check_aml_fs_subnets.py | 2 +- .../expansion_jobs_create_or_update.py | 49 + .../expansion_jobs_delete.py | 43 + .../generated_samples/expansion_jobs_get.py | 44 + .../expansion_jobs_list_by_aml_filesystem.py | 44 + .../expansion_jobs_update.py | 45 + .../get_required_aml_fs_subnets_size.py | 2 +- .../generated_samples/import_job_update.py | 2 +- .../import_jobs_create_or_update.py | 2 +- .../generated_samples/import_jobs_delete.py | 2 +- .../generated_samples/import_jobs_get.py | 2 +- .../import_jobs_list_by_aml_filesystem.py | 2 +- .../generated_samples/operations_list.py | 2 +- .../generated_samples/pause_priming_job.py | 2 +- .../generated_samples/resume_priming_job.py | 2 +- .../generated_samples/skus_list.py | 2 +- .../space_allocation_post.py | 2 +- .../generated_samples/start_priming_job.py | 2 +- .../generated_samples/stop_priming_job.py | 2 +- .../storage_targets_create_or_update.py | 2 +- ...orage_targets_create_or_update_blob_nfs.py | 2 +- ...e_targets_create_or_update_no_junctions.py | 2 +- .../storage_targets_delete.py | 2 +- .../storage_targets_dns_refresh.py | 2 +- .../storage_targets_flush.py | 2 +- .../generated_samples/storage_targets_get.py | 2 +- .../storage_targets_invalidate.py | 2 +- .../storage_targets_list_by_cache.py | 2 +- .../storage_targets_restore_defaults.py | 2 +- .../storage_targets_resume.py | 2 +- .../storage_targets_suspend.py | 2 +- .../generated_samples/usage_models_list.py | 2 +- .../test_storage_cache_management.py | 4 +- ...e_management_aml_filesystems_operations.py | 18 +- ...gement_aml_filesystems_operations_async.py | 18 +- ...he_management_asc_operations_operations.py | 2 +- ...agement_asc_operations_operations_async.py | 2 +- ..._cache_management_asc_usages_operations.py | 2 +- ..._management_asc_usages_operations_async.py | 2 +- .../test_storage_cache_management_async.py | 4 +- ..._management_auto_export_jobs_operations.py | 10 +- ...ement_auto_export_jobs_operations_async.py | 10 +- ..._management_auto_import_jobs_operations.py | 10 +- ...ement_auto_import_jobs_operations_async.py | 10 +- ...rage_cache_management_caches_operations.py | 32 +- ...ache_management_caches_operations_async.py | 32 +- ...he_management_expansion_jobs_operations.py | 107 ++ ...agement_expansion_jobs_operations_async.py | 114 ++ ...cache_management_import_jobs_operations.py | 10 +- ...management_import_jobs_operations_async.py | 10 +- ...est_storage_cache_management_operations.py | 2 +- ...orage_cache_management_operations_async.py | 2 +- ...torage_cache_management_skus_operations.py | 2 +- ..._cache_management_skus_operations_async.py | 2 +- ...he_management_storage_target_operations.py | 8 +- ...agement_storage_target_operations_async.py | 8 +- ...e_management_storage_targets_operations.py | 12 +- ...gement_storage_targets_operations_async.py | 12 +- ...ache_management_usage_models_operations.py | 2 +- ...anagement_usage_models_operations_async.py | 2 +- .../azure-mgmt-storagecache/pyproject.toml | 6 +- 136 files changed, 2916 insertions(+), 287 deletions(-) create mode 100644 sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_expansion_jobs_operations.py create mode 100644 sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_expansion_jobs_operations.py create mode 100644 sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_create_or_update.py create mode 100644 sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_delete.py create mode 100644 sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_get.py create mode 100644 sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_list_by_aml_filesystem.py create mode 100644 sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_update.py create mode 100644 sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_expansion_jobs_operations.py create mode 100644 sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_expansion_jobs_operations_async.py diff --git a/sdk/storage/azure-mgmt-storagecache/CHANGELOG.md b/sdk/storage/azure-mgmt-storagecache/CHANGELOG.md index 816be2d594d5..66ad9b67720b 100644 --- a/sdk/storage/azure-mgmt-storagecache/CHANGELOG.md +++ b/sdk/storage/azure-mgmt-storagecache/CHANGELOG.md @@ -1,5 +1,20 @@ # Release History +## 3.1.0 (2026-03-15) + +### Features Added + + - Client `StorageCacheManagementClient` added operation group `expansion_jobs` + - Model `AmlFilesystem` added property `current_storage_capacity_ti_b` + - Model `AmlFilesystem` added property `cluster_uuid` + - Enum `AmlFilesystemHealthStateType` added member `EXPANDING` + - Added model `ExpansionJob` + - Added enum `ExpansionJobPropertiesProvisioningState` + - Added enum `ExpansionJobStatusType` + - Added model `ExpansionJobUpdate` + - Added model `ExpansionJobsListResult` + - Added model `ExpansionJobsOperations` + ## 3.0.1 (2025-10-09) ### Bugs Fixed diff --git a/sdk/storage/azure-mgmt-storagecache/_metadata.json b/sdk/storage/azure-mgmt-storagecache/_metadata.json index 57bcbbe236a3..ef53ec78845c 100644 --- a/sdk/storage/azure-mgmt-storagecache/_metadata.json +++ b/sdk/storage/azure-mgmt-storagecache/_metadata.json @@ -1,11 +1,11 @@ { - "commit": "e976bdf4af8961332863ff779839f30fb2aec77a", + "commit": "3de705b1f6ad03adaae250fda48fae13de1f5b6a", "repository_url": "https://github.com/Azure/azure-rest-api-specs", "autorest": "3.10.2", "use": [ - "@autorest/python@6.39.0", + "@autorest/python@6.48.0", "@autorest/modelerfour@4.27.0" ], - "autorest_command": "autorest specification/storagecache/resource-manager/readme.md --generate-sample=True --generate-test=True --include-x-ms-examples-original-file=True --package-mode=azure-mgmt --python --python-sdks-folder=/mnt/vss/_work/1/azure-sdk-for-python/sdk --use=@autorest/python@6.39.0 --use=@autorest/modelerfour@4.27.0 --version=3.10.2 --version-tolerant=False", - "readme": "specification/storagecache/resource-manager/readme.md" + "autorest_command": "autorest specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/readme.md --generate-sample=True --generate-test=True --include-x-ms-examples-original-file=True --package-mode=azure-mgmt --python --python-sdks-folder=/mnt/vss/_work/1/azure-sdk-for-python/sdk --use=@autorest/python@6.48.0 --use=@autorest/modelerfour@4.27.0 --version=3.10.2 --version-tolerant=False", + "readme": "specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/readme.md" } \ No newline at end of file diff --git a/sdk/storage/azure-mgmt-storagecache/apiview-properties.json b/sdk/storage/azure-mgmt-storagecache/apiview-properties.json index f44717584e99..2fba3b7eae0f 100644 --- a/sdk/storage/azure-mgmt-storagecache/apiview-properties.json +++ b/sdk/storage/azure-mgmt-storagecache/apiview-properties.json @@ -57,6 +57,9 @@ "azure.mgmt.storagecache.models.ErrorAdditionalInfo": null, "azure.mgmt.storagecache.models.ErrorDetail": null, "azure.mgmt.storagecache.models.ErrorResponse": null, + "azure.mgmt.storagecache.models.ExpansionJob": null, + "azure.mgmt.storagecache.models.ExpansionJobsListResult": null, + "azure.mgmt.storagecache.models.ExpansionJobUpdate": null, "azure.mgmt.storagecache.models.ImportJob": null, "azure.mgmt.storagecache.models.ImportJobsListResult": null, "azure.mgmt.storagecache.models.ImportJobUpdate": null, @@ -110,6 +113,8 @@ "azure.mgmt.storagecache.models.AutoImportJobPropertiesProvisioningState": null, "azure.mgmt.storagecache.models.AutoImportJobPropertiesAdminStatus": null, "azure.mgmt.storagecache.models.AutoImportJobState": null, + "azure.mgmt.storagecache.models.ExpansionJobPropertiesProvisioningState": null, + "azure.mgmt.storagecache.models.ExpansionJobStatusType": null, "azure.mgmt.storagecache.models.MetricAggregationType": null, "azure.mgmt.storagecache.models.ReasonCode": null, "azure.mgmt.storagecache.models.NfsAccessRuleScope": null, @@ -172,6 +177,16 @@ "azure.mgmt.storagecache.aio.operations.AutoImportJobsOperations.begin_update": null, "azure.mgmt.storagecache.operations.AutoImportJobsOperations.list_by_aml_filesystem": null, "azure.mgmt.storagecache.aio.operations.AutoImportJobsOperations.list_by_aml_filesystem": null, + "azure.mgmt.storagecache.operations.ExpansionJobsOperations.begin_delete": null, + "azure.mgmt.storagecache.aio.operations.ExpansionJobsOperations.begin_delete": null, + "azure.mgmt.storagecache.operations.ExpansionJobsOperations.get": null, + "azure.mgmt.storagecache.aio.operations.ExpansionJobsOperations.get": null, + "azure.mgmt.storagecache.operations.ExpansionJobsOperations.begin_create_or_update": null, + "azure.mgmt.storagecache.aio.operations.ExpansionJobsOperations.begin_create_or_update": null, + "azure.mgmt.storagecache.operations.ExpansionJobsOperations.begin_update": null, + "azure.mgmt.storagecache.aio.operations.ExpansionJobsOperations.begin_update": null, + "azure.mgmt.storagecache.operations.ExpansionJobsOperations.list_by_aml_filesystem": null, + "azure.mgmt.storagecache.aio.operations.ExpansionJobsOperations.list_by_aml_filesystem": null, "azure.mgmt.storagecache.StorageCacheManagementClient.check_aml_fs_subnets": null, "azure.mgmt.storagecache.aio.StorageCacheManagementClient.check_aml_fs_subnets": null, "azure.mgmt.storagecache.StorageCacheManagementClient.get_required_aml_fs_subnets_size": null, diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/_configuration.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/_configuration.py index 82b58134b4d1..0369aa506057 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/_configuration.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/_configuration.py @@ -31,7 +31,7 @@ class StorageCacheManagementClientConfiguration: # pylint: disable=too-many-ins :param cloud_setting: The cloud setting for which to get the ARM endpoint. Default value is None. :type cloud_setting: ~azure.core.AzureClouds - :keyword api_version: Api Version. Default value is "2025-07-01". Note that overriding this + :keyword api_version: Api Version. Default value is "2026-01-01". Note that overriding this default value may result in unsupported behavior. :paramtype api_version: str """ @@ -43,7 +43,7 @@ def __init__( cloud_setting: Optional["AzureClouds"] = None, **kwargs: Any ) -> None: - api_version: str = kwargs.pop("api_version", "2025-07-01") + api_version: str = kwargs.pop("api_version", "2026-01-01") if credential is None: raise ValueError("Parameter 'credential' must not be None.") diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/_storage_cache_management_client.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/_storage_cache_management_client.py index 18b22574c2d1..507632dcbc3b 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/_storage_cache_management_client.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/_storage_cache_management_client.py @@ -27,6 +27,7 @@ AutoExportJobsOperations, AutoImportJobsOperations, CachesOperations, + ExpansionJobsOperations, ImportJobsOperations, Operations, SkusOperations, @@ -56,6 +57,8 @@ class StorageCacheManagementClient( :vartype import_jobs: azure.mgmt.storagecache.operations.ImportJobsOperations :ivar auto_import_jobs: AutoImportJobsOperations operations :vartype auto_import_jobs: azure.mgmt.storagecache.operations.AutoImportJobsOperations + :ivar expansion_jobs: ExpansionJobsOperations operations + :vartype expansion_jobs: azure.mgmt.storagecache.operations.ExpansionJobsOperations :ivar operations: Operations operations :vartype operations: azure.mgmt.storagecache.operations.Operations :ivar skus: SkusOperations operations @@ -81,7 +84,7 @@ class StorageCacheManagementClient( :keyword cloud_setting: The cloud setting for which to get the ARM endpoint. Default value is None. :paramtype cloud_setting: ~azure.core.AzureClouds - :keyword api_version: Api Version. Default value is "2025-07-01". Note that overriding this + :keyword api_version: Api Version. Default value is "2026-01-01". Note that overriding this default value may result in unsupported behavior. :paramtype api_version: str :keyword int polling_interval: Default waiting time between two polls for LRO operations if no @@ -138,6 +141,7 @@ def __init__( self.auto_export_jobs = AutoExportJobsOperations(self._client, self._config, self._serialize, self._deserialize) self.import_jobs = ImportJobsOperations(self._client, self._config, self._serialize, self._deserialize) self.auto_import_jobs = AutoImportJobsOperations(self._client, self._config, self._serialize, self._deserialize) + self.expansion_jobs = ExpansionJobsOperations(self._client, self._config, self._serialize, self._deserialize) self.operations = Operations(self._client, self._config, self._serialize, self._deserialize) self.skus = SkusOperations(self._client, self._config, self._serialize, self._deserialize) self.usage_models = UsageModelsOperations(self._client, self._config, self._serialize, self._deserialize) diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/_utils/serialization.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/_utils/serialization.py index 5f250836cf4b..6da830e0cf4a 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/_utils/serialization.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/_utils/serialization.py @@ -787,7 +787,7 @@ def serialize_data(self, data, data_type, **kwargs): # If dependencies is empty, try with current data class # It has to be a subclass of Enum anyway - enum_type = self.dependencies.get(data_type, data.__class__) + enum_type = self.dependencies.get(data_type, cast(type, data.__class__)) if issubclass(enum_type, Enum): return Serializer.serialize_enum(data, enum_obj=enum_type) @@ -821,13 +821,20 @@ def serialize_basic(cls, data, data_type, **kwargs): :param str data_type: Type of object in the iterable. :rtype: str, int, float, bool :return: serialized object + :raises TypeError: raise if data_type is not one of str, int, float, bool. """ custom_serializer = cls._get_custom_serializers(data_type, **kwargs) if custom_serializer: return custom_serializer(data) if data_type == "str": return cls.serialize_unicode(data) - return eval(data_type)(data) # nosec # pylint: disable=eval-used + if data_type == "int": + return int(data) + if data_type == "float": + return float(data) + if data_type == "bool": + return bool(data) + raise TypeError("Unknown basic data type: {}".format(data_type)) @classmethod def serialize_unicode(cls, data): @@ -1757,7 +1764,7 @@ def deserialize_basic(self, attr, data_type): # pylint: disable=too-many-return :param str data_type: deserialization data type. :return: Deserialized basic type. :rtype: str, int, float or bool - :raises TypeError: if string format is not valid. + :raises TypeError: if string format is not valid or data_type is not one of str, int, float, bool. """ # If we're here, data is supposed to be a basic type. # If it's still an XML node, take the text @@ -1783,7 +1790,11 @@ def deserialize_basic(self, attr, data_type): # pylint: disable=too-many-return if data_type == "str": return self.deserialize_unicode(attr) - return eval(data_type)(attr) # nosec # pylint: disable=eval-used + if data_type == "int": + return int(attr) + if data_type == "float": + return float(attr) + raise TypeError("Unknown basic data type: {}".format(data_type)) @staticmethod def deserialize_unicode(data): diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/_version.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/_version.py index 42cbaea39a1c..47babc28d5ed 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/_version.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/_version.py @@ -6,4 +6,4 @@ # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -VERSION = "3.0.1" +VERSION = "3.1.0" diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/_configuration.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/_configuration.py index 564ac5f6d3ac..7fdf8c55af53 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/_configuration.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/_configuration.py @@ -31,7 +31,7 @@ class StorageCacheManagementClientConfiguration: # pylint: disable=too-many-ins :param cloud_setting: The cloud setting for which to get the ARM endpoint. Default value is None. :type cloud_setting: ~azure.core.AzureClouds - :keyword api_version: Api Version. Default value is "2025-07-01". Note that overriding this + :keyword api_version: Api Version. Default value is "2026-01-01". Note that overriding this default value may result in unsupported behavior. :paramtype api_version: str """ @@ -43,7 +43,7 @@ def __init__( cloud_setting: Optional["AzureClouds"] = None, **kwargs: Any ) -> None: - api_version: str = kwargs.pop("api_version", "2025-07-01") + api_version: str = kwargs.pop("api_version", "2026-01-01") if credential is None: raise ValueError("Parameter 'credential' must not be None.") diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/_storage_cache_management_client.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/_storage_cache_management_client.py index f8c78f298266..b184919e3447 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/_storage_cache_management_client.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/_storage_cache_management_client.py @@ -27,6 +27,7 @@ AutoExportJobsOperations, AutoImportJobsOperations, CachesOperations, + ExpansionJobsOperations, ImportJobsOperations, Operations, SkusOperations, @@ -56,6 +57,8 @@ class StorageCacheManagementClient( :vartype import_jobs: azure.mgmt.storagecache.aio.operations.ImportJobsOperations :ivar auto_import_jobs: AutoImportJobsOperations operations :vartype auto_import_jobs: azure.mgmt.storagecache.aio.operations.AutoImportJobsOperations + :ivar expansion_jobs: ExpansionJobsOperations operations + :vartype expansion_jobs: azure.mgmt.storagecache.aio.operations.ExpansionJobsOperations :ivar operations: Operations operations :vartype operations: azure.mgmt.storagecache.aio.operations.Operations :ivar skus: SkusOperations operations @@ -81,7 +84,7 @@ class StorageCacheManagementClient( :keyword cloud_setting: The cloud setting for which to get the ARM endpoint. Default value is None. :paramtype cloud_setting: ~azure.core.AzureClouds - :keyword api_version: Api Version. Default value is "2025-07-01". Note that overriding this + :keyword api_version: Api Version. Default value is "2026-01-01". Note that overriding this default value may result in unsupported behavior. :paramtype api_version: str :keyword int polling_interval: Default waiting time between two polls for LRO operations if no @@ -140,6 +143,7 @@ def __init__( self.auto_export_jobs = AutoExportJobsOperations(self._client, self._config, self._serialize, self._deserialize) self.import_jobs = ImportJobsOperations(self._client, self._config, self._serialize, self._deserialize) self.auto_import_jobs = AutoImportJobsOperations(self._client, self._config, self._serialize, self._deserialize) + self.expansion_jobs = ExpansionJobsOperations(self._client, self._config, self._serialize, self._deserialize) self.operations = Operations(self._client, self._config, self._serialize, self._deserialize) self.skus = SkusOperations(self._client, self._config, self._serialize, self._deserialize) self.usage_models = UsageModelsOperations(self._client, self._config, self._serialize, self._deserialize) diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/__init__.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/__init__.py index 356469ae8c49..bf2a50f44406 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/__init__.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/__init__.py @@ -17,6 +17,7 @@ from ._auto_export_jobs_operations import AutoExportJobsOperations # type: ignore from ._import_jobs_operations import ImportJobsOperations # type: ignore from ._auto_import_jobs_operations import AutoImportJobsOperations # type: ignore +from ._expansion_jobs_operations import ExpansionJobsOperations # type: ignore from ._storage_cache_management_client_operations import _StorageCacheManagementClientOperationsMixin # type: ignore # pylint: disable=unused-import from ._operations import Operations # type: ignore from ._skus_operations import SkusOperations # type: ignore @@ -36,6 +37,7 @@ "AutoExportJobsOperations", "ImportJobsOperations", "AutoImportJobsOperations", + "ExpansionJobsOperations", "Operations", "SkusOperations", "UsageModelsOperations", diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_aml_filesystems_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_aml_filesystems_operations.py index 9877d046518b..50170db0a9c2 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_aml_filesystems_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_aml_filesystems_operations.py @@ -46,9 +46,9 @@ ) from .._configuration import StorageCacheManagementClientConfiguration -List = list T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, dict[str, Any]], Any]] +List = list class AmlFilesystemsOperations: diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_asc_operations_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_asc_operations_operations.py index 5434f9e940ed..6361199f5258 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_asc_operations_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_asc_operations_operations.py @@ -30,6 +30,7 @@ T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, dict[str, Any]], Any]] +List = list class AscOperationsOperations: diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_asc_usages_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_asc_usages_operations.py index 8d64a1c39581..ef18ef6e9a2e 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_asc_usages_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_asc_usages_operations.py @@ -30,9 +30,9 @@ from ...operations._asc_usages_operations import build_list_request from .._configuration import StorageCacheManagementClientConfiguration -List = list T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, dict[str, Any]], Any]] +List = list class AscUsagesOperations: diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_auto_export_jobs_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_auto_export_jobs_operations.py index f835dbd4dec4..d5deb485a6da 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_auto_export_jobs_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_auto_export_jobs_operations.py @@ -44,6 +44,7 @@ T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, dict[str, Any]], Any]] +List = list class AutoExportJobsOperations: @@ -109,7 +110,10 @@ async def _delete_initial( except (StreamConsumedError, StreamClosedError): pass map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) response_headers = {} @@ -241,7 +245,10 @@ async def get( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) deserialized = self._deserialize("AutoExportJob", pipeline_response.http_response) @@ -310,7 +317,10 @@ async def _create_or_update_initial( except (StreamConsumedError, StreamClosedError): pass map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) response_headers = {} @@ -535,7 +545,10 @@ async def _update_initial( except (StreamConsumedError, StreamClosedError): pass map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) response_headers = {} diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_auto_import_jobs_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_auto_import_jobs_operations.py index 33d301dbfb75..56249d01277a 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_auto_import_jobs_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_auto_import_jobs_operations.py @@ -44,6 +44,7 @@ T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, dict[str, Any]], Any]] +List = list class AutoImportJobsOperations: @@ -109,7 +110,10 @@ async def _delete_initial( except (StreamConsumedError, StreamClosedError): pass map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) response_headers = {} @@ -241,7 +245,10 @@ async def get( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) deserialized = self._deserialize("AutoImportJob", pipeline_response.http_response) @@ -310,7 +317,10 @@ async def _create_or_update_initial( except (StreamConsumedError, StreamClosedError): pass map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) response_headers = {} @@ -535,7 +545,10 @@ async def _update_initial( except (StreamConsumedError, StreamClosedError): pass map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) response_headers = {} diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_caches_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_caches_operations.py index b19ddbc38882..ed35932da23f 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_caches_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_caches_operations.py @@ -54,9 +54,9 @@ ) from .._configuration import StorageCacheManagementClientConfiguration -List = list T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, dict[str, Any]], Any]] +List = list class CachesOperations: # pylint: disable=too-many-public-methods diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_expansion_jobs_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_expansion_jobs_operations.py new file mode 100644 index 000000000000..da6ec8bc4ac1 --- /dev/null +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_expansion_jobs_operations.py @@ -0,0 +1,797 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- +from collections.abc import MutableMapping +from io import IOBase +from typing import Any, AsyncIterator, Callable, IO, Optional, TypeVar, Union, cast, overload +import urllib.parse + +from azure.core import AsyncPipelineClient +from azure.core.async_paging import AsyncItemPaged, AsyncList +from azure.core.exceptions import ( + ClientAuthenticationError, + HttpResponseError, + ResourceExistsError, + ResourceNotFoundError, + ResourceNotModifiedError, + StreamClosedError, + StreamConsumedError, + map_error, +) +from azure.core.pipeline import PipelineResponse +from azure.core.polling import AsyncLROPoller, AsyncNoPolling, AsyncPollingMethod +from azure.core.rest import AsyncHttpResponse, HttpRequest +from azure.core.tracing.decorator import distributed_trace +from azure.core.tracing.decorator_async import distributed_trace_async +from azure.core.utils import case_insensitive_dict +from azure.mgmt.core.exceptions import ARMErrorFormat +from azure.mgmt.core.polling.async_arm_polling import AsyncARMPolling + +from ... import models as _models +from ..._utils.serialization import Deserializer, Serializer +from ...operations._expansion_jobs_operations import ( + build_create_or_update_request, + build_delete_request, + build_get_request, + build_list_by_aml_filesystem_request, + build_update_request, +) +from .._configuration import StorageCacheManagementClientConfiguration + +T = TypeVar("T") +ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, dict[str, Any]], Any]] +List = list + + +class ExpansionJobsOperations: + """ + .. warning:: + **DO NOT** instantiate this class directly. + + Instead, you should access the following operations through + :class:`~azure.mgmt.storagecache.aio.StorageCacheManagementClient`'s + :attr:`expansion_jobs` attribute. + """ + + models = _models + + def __init__(self, *args, **kwargs) -> None: + input_args = list(args) + self._client: AsyncPipelineClient = input_args.pop(0) if input_args else kwargs.pop("client") + self._config: StorageCacheManagementClientConfiguration = ( + input_args.pop(0) if input_args else kwargs.pop("config") + ) + self._serialize: Serializer = input_args.pop(0) if input_args else kwargs.pop("serializer") + self._deserialize: Deserializer = input_args.pop(0) if input_args else kwargs.pop("deserializer") + + async def _delete_initial( + self, resource_group_name: str, aml_filesystem_name: str, expansion_job_name: str, **kwargs: Any + ) -> AsyncIterator[bytes]: + error_map: MutableMapping = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._config.api_version)) + cls: ClsType[AsyncIterator[bytes]] = kwargs.pop("cls", None) + + _request = build_delete_request( + resource_group_name=resource_group_name, + aml_filesystem_name=aml_filesystem_name, + expansion_job_name=expansion_job_name, + subscription_id=self._config.subscription_id, + api_version=api_version, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _decompress = kwargs.pop("decompress", True) + _stream = True + pipeline_response: PipelineResponse = await self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [202, 204]: + try: + await response.read() # Load the body in memory and close the socket + except (StreamConsumedError, StreamClosedError): + pass + map_error(status_code=response.status_code, response=response, error_map=error_map) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) + raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) + + response_headers = {} + if response.status_code == 202: + response_headers["Location"] = self._deserialize("str", response.headers.get("Location")) + response_headers["Azure-AsyncOperation"] = self._deserialize( + "str", response.headers.get("Azure-AsyncOperation") + ) + + deserialized = response.stream_download(self._client._pipeline, decompress=_decompress) + + if cls: + return cls(pipeline_response, deserialized, response_headers) # type: ignore + + return deserialized # type: ignore + + @distributed_trace_async + async def begin_delete( + self, resource_group_name: str, aml_filesystem_name: str, expansion_job_name: str, **kwargs: Any + ) -> AsyncLROPoller[None]: + """Schedules an expansion job for deletion. + + :param resource_group_name: The name of the resource group. The name is case insensitive. + Required. + :type resource_group_name: str + :param aml_filesystem_name: Name for the AML file system. Allows alphanumerics, underscores, + and hyphens. Start and end with alphanumeric. Required. + :type aml_filesystem_name: str + :param expansion_job_name: Name for the expansion job. Allows alphanumerics, underscores, and + hyphens. Start and end with alphanumeric. Required. + :type expansion_job_name: str + :return: An instance of AsyncLROPoller that returns either None or the result of cls(response) + :rtype: ~azure.core.polling.AsyncLROPoller[None] + :raises ~azure.core.exceptions.HttpResponseError: + """ + _headers = kwargs.pop("headers", {}) or {} + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._config.api_version)) + cls: ClsType[None] = kwargs.pop("cls", None) + polling: Union[bool, AsyncPollingMethod] = kwargs.pop("polling", True) + lro_delay = kwargs.pop("polling_interval", self._config.polling_interval) + cont_token: Optional[str] = kwargs.pop("continuation_token", None) + if cont_token is None: + raw_result = await self._delete_initial( + resource_group_name=resource_group_name, + aml_filesystem_name=aml_filesystem_name, + expansion_job_name=expansion_job_name, + api_version=api_version, + cls=lambda x, y, z: x, + headers=_headers, + params=_params, + **kwargs + ) + await raw_result.http_response.read() # type: ignore + kwargs.pop("error_map", None) + + def get_long_running_output(pipeline_response): # pylint: disable=inconsistent-return-statements + if cls: + return cls(pipeline_response, None, {}) # type: ignore + + if polling is True: + polling_method: AsyncPollingMethod = cast( + AsyncPollingMethod, AsyncARMPolling(lro_delay, lro_options={"final-state-via": "location"}, **kwargs) + ) + elif polling is False: + polling_method = cast(AsyncPollingMethod, AsyncNoPolling()) + else: + polling_method = polling + if cont_token: + return AsyncLROPoller[None].from_continuation_token( + polling_method=polling_method, + continuation_token=cont_token, + client=self._client, + deserialization_callback=get_long_running_output, + ) + return AsyncLROPoller[None](self._client, raw_result, get_long_running_output, polling_method) # type: ignore + + @distributed_trace_async + async def get( + self, resource_group_name: str, aml_filesystem_name: str, expansion_job_name: str, **kwargs: Any + ) -> _models.ExpansionJob: + """Returns an expansion job. + + :param resource_group_name: The name of the resource group. The name is case insensitive. + Required. + :type resource_group_name: str + :param aml_filesystem_name: Name for the AML file system. Allows alphanumerics, underscores, + and hyphens. Start and end with alphanumeric. Required. + :type aml_filesystem_name: str + :param expansion_job_name: Name for the expansion job. Allows alphanumerics, underscores, and + hyphens. Start and end with alphanumeric. Required. + :type expansion_job_name: str + :return: ExpansionJob or the result of cls(response) + :rtype: ~azure.mgmt.storagecache.models.ExpansionJob + :raises ~azure.core.exceptions.HttpResponseError: + """ + error_map: MutableMapping = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._config.api_version)) + cls: ClsType[_models.ExpansionJob] = kwargs.pop("cls", None) + + _request = build_get_request( + resource_group_name=resource_group_name, + aml_filesystem_name=aml_filesystem_name, + expansion_job_name=expansion_job_name, + subscription_id=self._config.subscription_id, + api_version=api_version, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = await self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [200]: + map_error(status_code=response.status_code, response=response, error_map=error_map) + raise HttpResponseError(response=response, error_format=ARMErrorFormat) + + deserialized = self._deserialize("ExpansionJob", pipeline_response.http_response) + + if cls: + return cls(pipeline_response, deserialized, {}) # type: ignore + + return deserialized # type: ignore + + async def _create_or_update_initial( + self, + resource_group_name: str, + aml_filesystem_name: str, + expansion_job_name: str, + expansion_job: Union[_models.ExpansionJob, IO[bytes]], + **kwargs: Any + ) -> AsyncIterator[bytes]: + error_map: MutableMapping = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._config.api_version)) + content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) + cls: ClsType[AsyncIterator[bytes]] = kwargs.pop("cls", None) + + content_type = content_type or "application/json" + _json = None + _content = None + if isinstance(expansion_job, (IOBase, bytes)): + _content = expansion_job + else: + _json = self._serialize.body(expansion_job, "ExpansionJob") + + _request = build_create_or_update_request( + resource_group_name=resource_group_name, + aml_filesystem_name=aml_filesystem_name, + expansion_job_name=expansion_job_name, + subscription_id=self._config.subscription_id, + api_version=api_version, + content_type=content_type, + json=_json, + content=_content, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _decompress = kwargs.pop("decompress", True) + _stream = True + pipeline_response: PipelineResponse = await self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 201]: + try: + await response.read() # Load the body in memory and close the socket + except (StreamConsumedError, StreamClosedError): + pass + map_error(status_code=response.status_code, response=response, error_map=error_map) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) + raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) + + response_headers = {} + if response.status_code == 201: + response_headers["Location"] = self._deserialize("str", response.headers.get("Location")) + response_headers["Azure-AsyncOperation"] = self._deserialize( + "str", response.headers.get("Azure-AsyncOperation") + ) + + deserialized = response.stream_download(self._client._pipeline, decompress=_decompress) + + if cls: + return cls(pipeline_response, deserialized, response_headers) # type: ignore + + return deserialized # type: ignore + + @overload + async def begin_create_or_update( + self, + resource_group_name: str, + aml_filesystem_name: str, + expansion_job_name: str, + expansion_job: _models.ExpansionJob, + *, + content_type: str = "application/json", + **kwargs: Any + ) -> AsyncLROPoller[_models.ExpansionJob]: + """Create or update an expansion job. + + :param resource_group_name: The name of the resource group. The name is case insensitive. + Required. + :type resource_group_name: str + :param aml_filesystem_name: Name for the AML file system. Allows alphanumerics, underscores, + and hyphens. Start and end with alphanumeric. Required. + :type aml_filesystem_name: str + :param expansion_job_name: Name for the expansion job. Allows alphanumerics, underscores, and + hyphens. Start and end with alphanumeric. Required. + :type expansion_job_name: str + :param expansion_job: Object containing the user-selectable properties of the expansion job. If + read-only properties are included, they must match the existing values of those properties. + Required. + :type expansion_job: ~azure.mgmt.storagecache.models.ExpansionJob + :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. + Default value is "application/json". + :paramtype content_type: str + :return: An instance of AsyncLROPoller that returns either ExpansionJob or the result of + cls(response) + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.mgmt.storagecache.models.ExpansionJob] + :raises ~azure.core.exceptions.HttpResponseError: + """ + + @overload + async def begin_create_or_update( + self, + resource_group_name: str, + aml_filesystem_name: str, + expansion_job_name: str, + expansion_job: IO[bytes], + *, + content_type: str = "application/json", + **kwargs: Any + ) -> AsyncLROPoller[_models.ExpansionJob]: + """Create or update an expansion job. + + :param resource_group_name: The name of the resource group. The name is case insensitive. + Required. + :type resource_group_name: str + :param aml_filesystem_name: Name for the AML file system. Allows alphanumerics, underscores, + and hyphens. Start and end with alphanumeric. Required. + :type aml_filesystem_name: str + :param expansion_job_name: Name for the expansion job. Allows alphanumerics, underscores, and + hyphens. Start and end with alphanumeric. Required. + :type expansion_job_name: str + :param expansion_job: Object containing the user-selectable properties of the expansion job. If + read-only properties are included, they must match the existing values of those properties. + Required. + :type expansion_job: IO[bytes] + :keyword content_type: Body Parameter content-type. Content type parameter for binary body. + Default value is "application/json". + :paramtype content_type: str + :return: An instance of AsyncLROPoller that returns either ExpansionJob or the result of + cls(response) + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.mgmt.storagecache.models.ExpansionJob] + :raises ~azure.core.exceptions.HttpResponseError: + """ + + @distributed_trace_async + async def begin_create_or_update( + self, + resource_group_name: str, + aml_filesystem_name: str, + expansion_job_name: str, + expansion_job: Union[_models.ExpansionJob, IO[bytes]], + **kwargs: Any + ) -> AsyncLROPoller[_models.ExpansionJob]: + """Create or update an expansion job. + + :param resource_group_name: The name of the resource group. The name is case insensitive. + Required. + :type resource_group_name: str + :param aml_filesystem_name: Name for the AML file system. Allows alphanumerics, underscores, + and hyphens. Start and end with alphanumeric. Required. + :type aml_filesystem_name: str + :param expansion_job_name: Name for the expansion job. Allows alphanumerics, underscores, and + hyphens. Start and end with alphanumeric. Required. + :type expansion_job_name: str + :param expansion_job: Object containing the user-selectable properties of the expansion job. If + read-only properties are included, they must match the existing values of those properties. Is + either a ExpansionJob type or a IO[bytes] type. Required. + :type expansion_job: ~azure.mgmt.storagecache.models.ExpansionJob or IO[bytes] + :return: An instance of AsyncLROPoller that returns either ExpansionJob or the result of + cls(response) + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.mgmt.storagecache.models.ExpansionJob] + :raises ~azure.core.exceptions.HttpResponseError: + """ + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._config.api_version)) + content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) + cls: ClsType[_models.ExpansionJob] = kwargs.pop("cls", None) + polling: Union[bool, AsyncPollingMethod] = kwargs.pop("polling", True) + lro_delay = kwargs.pop("polling_interval", self._config.polling_interval) + cont_token: Optional[str] = kwargs.pop("continuation_token", None) + if cont_token is None: + raw_result = await self._create_or_update_initial( + resource_group_name=resource_group_name, + aml_filesystem_name=aml_filesystem_name, + expansion_job_name=expansion_job_name, + expansion_job=expansion_job, + api_version=api_version, + content_type=content_type, + cls=lambda x, y, z: x, + headers=_headers, + params=_params, + **kwargs + ) + await raw_result.http_response.read() # type: ignore + kwargs.pop("error_map", None) + + def get_long_running_output(pipeline_response): + deserialized = self._deserialize("ExpansionJob", pipeline_response.http_response) + if cls: + return cls(pipeline_response, deserialized, {}) # type: ignore + return deserialized + + if polling is True: + polling_method: AsyncPollingMethod = cast( + AsyncPollingMethod, + AsyncARMPolling(lro_delay, lro_options={"final-state-via": "azure-async-operation"}, **kwargs), + ) + elif polling is False: + polling_method = cast(AsyncPollingMethod, AsyncNoPolling()) + else: + polling_method = polling + if cont_token: + return AsyncLROPoller[_models.ExpansionJob].from_continuation_token( + polling_method=polling_method, + continuation_token=cont_token, + client=self._client, + deserialization_callback=get_long_running_output, + ) + return AsyncLROPoller[_models.ExpansionJob]( + self._client, raw_result, get_long_running_output, polling_method # type: ignore + ) + + async def _update_initial( + self, + resource_group_name: str, + aml_filesystem_name: str, + expansion_job_name: str, + expansion_job: Union[_models.ExpansionJobUpdate, IO[bytes]], + **kwargs: Any + ) -> AsyncIterator[bytes]: + error_map: MutableMapping = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._config.api_version)) + content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) + cls: ClsType[AsyncIterator[bytes]] = kwargs.pop("cls", None) + + content_type = content_type or "application/json" + _json = None + _content = None + if isinstance(expansion_job, (IOBase, bytes)): + _content = expansion_job + else: + _json = self._serialize.body(expansion_job, "ExpansionJobUpdate") + + _request = build_update_request( + resource_group_name=resource_group_name, + aml_filesystem_name=aml_filesystem_name, + expansion_job_name=expansion_job_name, + subscription_id=self._config.subscription_id, + api_version=api_version, + content_type=content_type, + json=_json, + content=_content, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _decompress = kwargs.pop("decompress", True) + _stream = True + pipeline_response: PipelineResponse = await self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 202]: + try: + await response.read() # Load the body in memory and close the socket + except (StreamConsumedError, StreamClosedError): + pass + map_error(status_code=response.status_code, response=response, error_map=error_map) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) + raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) + + response_headers = {} + if response.status_code == 202: + response_headers["Location"] = self._deserialize("str", response.headers.get("Location")) + response_headers["Azure-AsyncOperation"] = self._deserialize( + "str", response.headers.get("Azure-AsyncOperation") + ) + + deserialized = response.stream_download(self._client._pipeline, decompress=_decompress) + + if cls: + return cls(pipeline_response, deserialized, response_headers) # type: ignore + + return deserialized # type: ignore + + @overload + async def begin_update( + self, + resource_group_name: str, + aml_filesystem_name: str, + expansion_job_name: str, + expansion_job: _models.ExpansionJobUpdate, + *, + content_type: str = "application/json", + **kwargs: Any + ) -> AsyncLROPoller[_models.ExpansionJob]: + """Update an expansion job instance. + + :param resource_group_name: The name of the resource group. The name is case insensitive. + Required. + :type resource_group_name: str + :param aml_filesystem_name: Name for the AML file system. Allows alphanumerics, underscores, + and hyphens. Start and end with alphanumeric. Required. + :type aml_filesystem_name: str + :param expansion_job_name: Name for the expansion job. Allows alphanumerics, underscores, and + hyphens. Start and end with alphanumeric. Required. + :type expansion_job_name: str + :param expansion_job: Object containing the user-selectable properties of the expansion job. If + read-only properties are included, they must match the existing values of those properties. + Required. + :type expansion_job: ~azure.mgmt.storagecache.models.ExpansionJobUpdate + :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. + Default value is "application/json". + :paramtype content_type: str + :return: An instance of AsyncLROPoller that returns either ExpansionJob or the result of + cls(response) + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.mgmt.storagecache.models.ExpansionJob] + :raises ~azure.core.exceptions.HttpResponseError: + """ + + @overload + async def begin_update( + self, + resource_group_name: str, + aml_filesystem_name: str, + expansion_job_name: str, + expansion_job: IO[bytes], + *, + content_type: str = "application/json", + **kwargs: Any + ) -> AsyncLROPoller[_models.ExpansionJob]: + """Update an expansion job instance. + + :param resource_group_name: The name of the resource group. The name is case insensitive. + Required. + :type resource_group_name: str + :param aml_filesystem_name: Name for the AML file system. Allows alphanumerics, underscores, + and hyphens. Start and end with alphanumeric. Required. + :type aml_filesystem_name: str + :param expansion_job_name: Name for the expansion job. Allows alphanumerics, underscores, and + hyphens. Start and end with alphanumeric. Required. + :type expansion_job_name: str + :param expansion_job: Object containing the user-selectable properties of the expansion job. If + read-only properties are included, they must match the existing values of those properties. + Required. + :type expansion_job: IO[bytes] + :keyword content_type: Body Parameter content-type. Content type parameter for binary body. + Default value is "application/json". + :paramtype content_type: str + :return: An instance of AsyncLROPoller that returns either ExpansionJob or the result of + cls(response) + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.mgmt.storagecache.models.ExpansionJob] + :raises ~azure.core.exceptions.HttpResponseError: + """ + + @distributed_trace_async + async def begin_update( + self, + resource_group_name: str, + aml_filesystem_name: str, + expansion_job_name: str, + expansion_job: Union[_models.ExpansionJobUpdate, IO[bytes]], + **kwargs: Any + ) -> AsyncLROPoller[_models.ExpansionJob]: + """Update an expansion job instance. + + :param resource_group_name: The name of the resource group. The name is case insensitive. + Required. + :type resource_group_name: str + :param aml_filesystem_name: Name for the AML file system. Allows alphanumerics, underscores, + and hyphens. Start and end with alphanumeric. Required. + :type aml_filesystem_name: str + :param expansion_job_name: Name for the expansion job. Allows alphanumerics, underscores, and + hyphens. Start and end with alphanumeric. Required. + :type expansion_job_name: str + :param expansion_job: Object containing the user-selectable properties of the expansion job. If + read-only properties are included, they must match the existing values of those properties. Is + either a ExpansionJobUpdate type or a IO[bytes] type. Required. + :type expansion_job: ~azure.mgmt.storagecache.models.ExpansionJobUpdate or IO[bytes] + :return: An instance of AsyncLROPoller that returns either ExpansionJob or the result of + cls(response) + :rtype: ~azure.core.polling.AsyncLROPoller[~azure.mgmt.storagecache.models.ExpansionJob] + :raises ~azure.core.exceptions.HttpResponseError: + """ + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._config.api_version)) + content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) + cls: ClsType[_models.ExpansionJob] = kwargs.pop("cls", None) + polling: Union[bool, AsyncPollingMethod] = kwargs.pop("polling", True) + lro_delay = kwargs.pop("polling_interval", self._config.polling_interval) + cont_token: Optional[str] = kwargs.pop("continuation_token", None) + if cont_token is None: + raw_result = await self._update_initial( + resource_group_name=resource_group_name, + aml_filesystem_name=aml_filesystem_name, + expansion_job_name=expansion_job_name, + expansion_job=expansion_job, + api_version=api_version, + content_type=content_type, + cls=lambda x, y, z: x, + headers=_headers, + params=_params, + **kwargs + ) + await raw_result.http_response.read() # type: ignore + kwargs.pop("error_map", None) + + def get_long_running_output(pipeline_response): + deserialized = self._deserialize("ExpansionJob", pipeline_response.http_response) + if cls: + return cls(pipeline_response, deserialized, {}) # type: ignore + return deserialized + + if polling is True: + polling_method: AsyncPollingMethod = cast( + AsyncPollingMethod, + AsyncARMPolling(lro_delay, lro_options={"final-state-via": "azure-async-operation"}, **kwargs), + ) + elif polling is False: + polling_method = cast(AsyncPollingMethod, AsyncNoPolling()) + else: + polling_method = polling + if cont_token: + return AsyncLROPoller[_models.ExpansionJob].from_continuation_token( + polling_method=polling_method, + continuation_token=cont_token, + client=self._client, + deserialization_callback=get_long_running_output, + ) + return AsyncLROPoller[_models.ExpansionJob]( + self._client, raw_result, get_long_running_output, polling_method # type: ignore + ) + + @distributed_trace + def list_by_aml_filesystem( + self, resource_group_name: str, aml_filesystem_name: str, **kwargs: Any + ) -> AsyncItemPaged["_models.ExpansionJob"]: + """Returns all the expansion jobs the user has access to under an AML File System. + + :param resource_group_name: The name of the resource group. The name is case insensitive. + Required. + :type resource_group_name: str + :param aml_filesystem_name: Name for the AML file system. Allows alphanumerics, underscores, + and hyphens. Start and end with alphanumeric. Required. + :type aml_filesystem_name: str + :return: An iterator like instance of either ExpansionJob or the result of cls(response) + :rtype: ~azure.core.async_paging.AsyncItemPaged[~azure.mgmt.storagecache.models.ExpansionJob] + :raises ~azure.core.exceptions.HttpResponseError: + """ + _headers = kwargs.pop("headers", {}) or {} + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._config.api_version)) + cls: ClsType[_models.ExpansionJobsListResult] = kwargs.pop("cls", None) + + error_map: MutableMapping = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + def prepare_request(next_link=None): + if not next_link: + + _request = build_list_by_aml_filesystem_request( + resource_group_name=resource_group_name, + aml_filesystem_name=aml_filesystem_name, + subscription_id=self._config.subscription_id, + api_version=api_version, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + else: + # make call to next link with the client's api-version + _parsed_next_link = urllib.parse.urlparse(next_link) + _next_request_params = case_insensitive_dict( + { + key: [urllib.parse.quote(v) for v in value] + for key, value in urllib.parse.parse_qs(_parsed_next_link.query).items() + } + ) + _next_request_params["api-version"] = self._config.api_version + _request = HttpRequest( + "GET", urllib.parse.urljoin(next_link, _parsed_next_link.path), params=_next_request_params + ) + _request.url = self._client.format_url(_request.url) + _request.method = "GET" + return _request + + async def extract_data(pipeline_response): + deserialized = self._deserialize("ExpansionJobsListResult", pipeline_response) + list_of_elem = deserialized.value + if cls: + list_of_elem = cls(list_of_elem) # type: ignore + return deserialized.next_link or None, AsyncList(list_of_elem) + + async def get_next(next_link=None): + _request = prepare_request(next_link) + + _stream = False + pipeline_response: PipelineResponse = await self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + response = pipeline_response.http_response + + if response.status_code not in [200]: + map_error(status_code=response.status_code, response=response, error_map=error_map) + raise HttpResponseError(response=response, error_format=ARMErrorFormat) + + return pipeline_response + + return AsyncItemPaged(get_next, extract_data) diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_import_jobs_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_import_jobs_operations.py index 0e27de4b255c..22c0075a11c1 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_import_jobs_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_import_jobs_operations.py @@ -44,6 +44,7 @@ T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, dict[str, Any]], Any]] +List = list class ImportJobsOperations: @@ -109,7 +110,10 @@ async def _delete_initial( except (StreamConsumedError, StreamClosedError): pass map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) response_headers = {} @@ -241,7 +245,10 @@ async def get( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) deserialized = self._deserialize("ImportJob", pipeline_response.http_response) @@ -310,7 +317,10 @@ async def _create_or_update_initial( except (StreamConsumedError, StreamClosedError): pass map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) response_headers = {} @@ -535,7 +545,10 @@ async def _update_initial( except (StreamConsumedError, StreamClosedError): pass map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) response_headers = {} diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_operations.py index 291e7e1923c4..77f548e4d1d7 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_operations.py @@ -30,9 +30,9 @@ from ...operations._operations import build_list_request from .._configuration import StorageCacheManagementClientConfiguration -List = list T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, dict[str, Any]], Any]] +List = list class Operations: diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_skus_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_skus_operations.py index cf70f77b0966..9b9abe3dea92 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_skus_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_skus_operations.py @@ -30,9 +30,9 @@ from ...operations._skus_operations import build_list_request from .._configuration import StorageCacheManagementClientConfiguration -List = list T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, dict[str, Any]], Any]] +List = list class SkusOperations: diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_storage_cache_management_client_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_storage_cache_management_client_operations.py index b2ed2f7f93b3..cbf2dae9c276 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_storage_cache_management_client_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_storage_cache_management_client_operations.py @@ -35,6 +35,7 @@ T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, dict[str, Any]], Any]] +List = list class _StorageCacheManagementClientOperationsMixin( @@ -148,7 +149,10 @@ async def check_aml_fs_subnets( map_error(status_code=response.status_code, response=response, error_map=error_map) error = None if response.status_code == 400: - error = self._deserialize.failsafe_deserialize(_models.AmlFilesystemCheckSubnetError, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.AmlFilesystemCheckSubnetError, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) if cls: diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_storage_target_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_storage_target_operations.py index 345e034cb960..d40f94cdfe12 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_storage_target_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_storage_target_operations.py @@ -39,6 +39,7 @@ T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, dict[str, Any]], Any]] +List = list class StorageTargetOperations: diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_storage_targets_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_storage_targets_operations.py index 03697a11b683..51da5382cde4 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_storage_targets_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_storage_targets_operations.py @@ -45,6 +45,7 @@ T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, dict[str, Any]], Any]] +List = list class StorageTargetsOperations: diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_usage_models_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_usage_models_operations.py index 65d1ac703672..e13a7869c8f4 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_usage_models_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/aio/operations/_usage_models_operations.py @@ -30,9 +30,9 @@ from ...operations._usage_models_operations import build_list_request from .._configuration import StorageCacheManagementClientConfiguration -List = list T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, AsyncHttpResponse], T, dict[str, Any]], Any]] +List = list class UsageModelsOperations: diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/models/__init__.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/models/__init__.py index bdb007863545..f568ba86c994 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/models/__init__.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/models/__init__.py @@ -68,6 +68,9 @@ ErrorAdditionalInfo, ErrorDetail, ErrorResponse, + ExpansionJob, + ExpansionJobUpdate, + ExpansionJobsListResult, ImportJob, ImportJobUpdate, ImportJobsListResult, @@ -125,6 +128,8 @@ ConflictResolutionMode, CreatedByType, DomainJoinedType, + ExpansionJobPropertiesProvisioningState, + ExpansionJobStatusType, FilesystemSubnetStatusType, FirmwareStatusType, HealthStateType, @@ -202,6 +207,9 @@ "ErrorAdditionalInfo", "ErrorDetail", "ErrorResponse", + "ExpansionJob", + "ExpansionJobUpdate", + "ExpansionJobsListResult", "ImportJob", "ImportJobUpdate", "ImportJobsListResult", @@ -256,6 +264,8 @@ "ConflictResolutionMode", "CreatedByType", "DomainJoinedType", + "ExpansionJobPropertiesProvisioningState", + "ExpansionJobStatusType", "FilesystemSubnetStatusType", "FirmwareStatusType", "HealthStateType", diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/models/_models_py3.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/models/_models_py3.py index 027101c407f3..06f8db80efba 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/models/_models_py3.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/models/_models_py3.py @@ -145,6 +145,11 @@ class AmlFilesystem(TrackedResource): :vartype zones: list[str] :ivar storage_capacity_ti_b: The size of the AML file system, in TiB. This might be rounded up. :vartype storage_capacity_ti_b: float + :ivar current_storage_capacity_ti_b: The current storage capacity of the AML file system, in + TiB. This reflects the actual capacity including any expansions. + :vartype current_storage_capacity_ti_b: float + :ivar cluster_uuid: The unique identifier of the AML file system cluster. + :vartype cluster_uuid: str :ivar health: Health of the AML file system. :vartype health: ~azure.mgmt.storagecache.models.AmlFilesystemHealth :ivar provisioning_state: ARM provisioning state. Known values are: "Succeeded", "Failed", @@ -176,6 +181,8 @@ class AmlFilesystem(TrackedResource): "type": {"readonly": True}, "system_data": {"readonly": True}, "location": {"required": True}, + "current_storage_capacity_ti_b": {"readonly": True}, + "cluster_uuid": {"readonly": True}, "health": {"readonly": True}, "provisioning_state": {"readonly": True}, "client_info": {"readonly": True}, @@ -193,6 +200,8 @@ class AmlFilesystem(TrackedResource): "sku": {"key": "sku", "type": "SkuName"}, "zones": {"key": "zones", "type": "[str]"}, "storage_capacity_ti_b": {"key": "properties.storageCapacityTiB", "type": "float"}, + "current_storage_capacity_ti_b": {"key": "properties.currentStorageCapacityTiB", "type": "float"}, + "cluster_uuid": {"key": "properties.clusterUuid", "type": "str"}, "health": {"key": "properties.health", "type": "AmlFilesystemHealth"}, "provisioning_state": {"key": "properties.provisioningState", "type": "str"}, "filesystem_subnet": {"key": "properties.filesystemSubnet", "type": "str"}, @@ -257,6 +266,8 @@ def __init__( self.sku = sku self.zones = zones self.storage_capacity_ti_b = storage_capacity_ti_b + self.current_storage_capacity_ti_b: Optional[float] = None + self.cluster_uuid: Optional[str] = None self.health: Optional["_models.AmlFilesystemHealth"] = None self.provisioning_state: Optional[Union[str, "_models.AmlFilesystemProvisioningStateType"]] = None self.filesystem_subnet = filesystem_subnet @@ -533,7 +544,7 @@ class AmlFilesystemHealth(_serialization.Model): related to provisioning. :ivar state: List of AML file system health states. Known values are: "Unavailable", - "Available", "Degraded", "Transitioning", and "Maintenance". + "Available", "Degraded", "Transitioning", "Maintenance", and "Expanding". :vartype state: str or ~azure.mgmt.storagecache.models.AmlFilesystemHealthStateType :ivar status_code: Server-defined error code for the AML file system health. :vartype status_code: str @@ -557,7 +568,7 @@ def __init__( ) -> None: """ :keyword state: List of AML file system health states. Known values are: "Unavailable", - "Available", "Degraded", "Transitioning", and "Maintenance". + "Available", "Degraded", "Transitioning", "Maintenance", and "Expanding". :paramtype state: str or ~azure.mgmt.storagecache.models.AmlFilesystemHealthStateType :keyword status_code: Server-defined error code for the AML file system health. :paramtype status_code: str @@ -1393,7 +1404,7 @@ class AutoExportJob(TrackedResource): "last_completion_time_utc": {"key": "properties.status.lastCompletionTimeUTC", "type": "iso-8601"}, } - def __init__( # pylint: disable=too-many-locals + def __init__( self, *, location: str, @@ -2978,6 +2989,166 @@ def __init__(self, *, error: Optional["_models.ErrorDetail"] = None, **kwargs: A self.error = error +class ExpansionJob(TrackedResource): + """An expansion job instance. Follows Azure Resource Manager standards: + https://github.com/Azure/azure-resource-manager-rpc/blob/master/v1.0/resource-api-reference.md. + + Variables are only populated by the server, and will be ignored when sending a request. + + All required parameters must be populated in order to send to server. + + :ivar id: Fully qualified resource ID for the resource. Ex - + /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/{resourceProviderNamespace}/{resourceType}/{resourceName}. + :vartype id: str + :ivar name: The name of the resource. + :vartype name: str + :ivar type: The type of the resource. E.g. "Microsoft.Compute/virtualMachines" or + "Microsoft.Storage/storageAccounts". + :vartype type: str + :ivar system_data: Azure Resource Manager metadata containing createdBy and modifiedBy + information. + :vartype system_data: ~azure.mgmt.storagecache.models.SystemData + :ivar tags: Resource tags. + :vartype tags: dict[str, str] + :ivar location: The geo-location where the resource lives. Required. + :vartype location: str + :ivar provisioning_state: ARM provisioning state, see + https://github.com/Azure/azure-resource-manager-rpc/blob/master/v1.0/Addendum.md#provisioningstate-property. + Known values are: "Succeeded", "Failed", "Creating", "Deleting", "Updating", and "Canceled". + :vartype provisioning_state: str or + ~azure.mgmt.storagecache.models.ExpansionJobPropertiesProvisioningState + :ivar new_storage_capacity_ti_b: The new storage capacity in TiB for the AML file system after + expansion. This must be a multiple of the Sku step size, and greater than the current storage + capacity of the AML file system. + :vartype new_storage_capacity_ti_b: float + :ivar state: The operational state of the expansion job. InProgress indicates the expansion is + still running. Completed indicates expansion finished successfully. Failed means the expansion + was unable to complete due to a fatal error. Deleting indicates the expansion is being rolled + back. Known values are: "InProgress", "Completed", "Failed", "Deleting", and "RollingBack". + :vartype state: str or ~azure.mgmt.storagecache.models.ExpansionJobStatusType + :ivar status_code: Server-defined status code for expansion job. + :vartype status_code: str + :ivar status_message: Server-defined status message for expansion job. + :vartype status_message: str + :ivar percent_complete: The percentage of expansion job completion. + :vartype percent_complete: float + :ivar start_time_utc: The time (in UTC) the expansion job started. + :vartype start_time_utc: ~datetime.datetime + :ivar completion_time_utc: The time (in UTC) when the expansion job completed. Only populated + when job reaches a terminal state. + :vartype completion_time_utc: ~datetime.datetime + """ + + _validation = { + "id": {"readonly": True}, + "name": {"readonly": True}, + "type": {"readonly": True}, + "system_data": {"readonly": True}, + "location": {"required": True}, + "provisioning_state": {"readonly": True}, + "state": {"readonly": True}, + "status_code": {"readonly": True}, + "status_message": {"readonly": True}, + "percent_complete": {"readonly": True, "maximum": 100, "minimum": 0}, + "start_time_utc": {"readonly": True}, + "completion_time_utc": {"readonly": True}, + } + + _attribute_map = { + "id": {"key": "id", "type": "str"}, + "name": {"key": "name", "type": "str"}, + "type": {"key": "type", "type": "str"}, + "system_data": {"key": "systemData", "type": "SystemData"}, + "tags": {"key": "tags", "type": "{str}"}, + "location": {"key": "location", "type": "str"}, + "provisioning_state": {"key": "properties.provisioningState", "type": "str"}, + "new_storage_capacity_ti_b": {"key": "properties.newStorageCapacityTiB", "type": "float"}, + "state": {"key": "properties.status.state", "type": "str"}, + "status_code": {"key": "properties.status.statusCode", "type": "str"}, + "status_message": {"key": "properties.status.statusMessage", "type": "str"}, + "percent_complete": {"key": "properties.status.percentComplete", "type": "float"}, + "start_time_utc": {"key": "properties.status.startTimeUTC", "type": "iso-8601"}, + "completion_time_utc": {"key": "properties.status.completionTimeUTC", "type": "iso-8601"}, + } + + def __init__( + self, + *, + location: str, + tags: Optional[dict[str, str]] = None, + new_storage_capacity_ti_b: Optional[float] = None, + **kwargs: Any + ) -> None: + """ + :keyword tags: Resource tags. + :paramtype tags: dict[str, str] + :keyword location: The geo-location where the resource lives. Required. + :paramtype location: str + :keyword new_storage_capacity_ti_b: The new storage capacity in TiB for the AML file system + after expansion. This must be a multiple of the Sku step size, and greater than the current + storage capacity of the AML file system. + :paramtype new_storage_capacity_ti_b: float + """ + super().__init__(tags=tags, location=location, **kwargs) + self.provisioning_state: Optional[Union[str, "_models.ExpansionJobPropertiesProvisioningState"]] = None + self.new_storage_capacity_ti_b = new_storage_capacity_ti_b + self.state: Optional[Union[str, "_models.ExpansionJobStatusType"]] = None + self.status_code: Optional[str] = None + self.status_message: Optional[str] = None + self.percent_complete: Optional[float] = None + self.start_time_utc: Optional[datetime.datetime] = None + self.completion_time_utc: Optional[datetime.datetime] = None + + +class ExpansionJobsListResult(_serialization.Model): + """Result of the request to list expansion jobs. It contains a list of expansion jobs and a URL + link to get the next set of results. + + :ivar value: List of expansion jobs. + :vartype value: list[~azure.mgmt.storagecache.models.ExpansionJob] + :ivar next_link: URL to get the next set of expansion job list results, if there are any. + :vartype next_link: str + """ + + _attribute_map = { + "value": {"key": "value", "type": "[ExpansionJob]"}, + "next_link": {"key": "nextLink", "type": "str"}, + } + + def __init__( + self, *, value: Optional[list["_models.ExpansionJob"]] = None, next_link: Optional[str] = None, **kwargs: Any + ) -> None: + """ + :keyword value: List of expansion jobs. + :paramtype value: list[~azure.mgmt.storagecache.models.ExpansionJob] + :keyword next_link: URL to get the next set of expansion job list results, if there are any. + :paramtype next_link: str + """ + super().__init__(**kwargs) + self.value = value + self.next_link = next_link + + +class ExpansionJobUpdate(_serialization.Model): + """An expansion job update instance. + + :ivar tags: Resource tags. + :vartype tags: dict[str, str] + """ + + _attribute_map = { + "tags": {"key": "tags", "type": "{str}"}, + } + + def __init__(self, *, tags: Optional[dict[str, str]] = None, **kwargs: Any) -> None: + """ + :keyword tags: Resource tags. + :paramtype tags: dict[str, str] + """ + super().__init__(**kwargs) + self.tags = tags + + class ImportJob(TrackedResource): """An import job instance. Follows Azure Resource Manager standards: https://github.com/Azure/azure-resource-manager-rpc/blob/master/v1.0/resource-api-reference.md. diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/models/_storage_cache_management_client_enums.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/models/_storage_cache_management_client_enums.py index b33616407aab..897ed8204de6 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/models/_storage_cache_management_client_enums.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/models/_storage_cache_management_client_enums.py @@ -18,6 +18,7 @@ class AmlFilesystemHealthStateType(str, Enum, metaclass=CaseInsensitiveEnumMeta) DEGRADED = "Degraded" TRANSITIONING = "Transitioning" MAINTENANCE = "Maintenance" + EXPANDING = "Expanding" class AmlFilesystemIdentityType(str, Enum, metaclass=CaseInsensitiveEnumMeta): @@ -181,6 +182,32 @@ class DomainJoinedType(str, Enum, metaclass=CaseInsensitiveEnumMeta): ERROR = "Error" +class ExpansionJobPropertiesProvisioningState(str, Enum, metaclass=CaseInsensitiveEnumMeta): + """ARM provisioning state, see + https://github.com/Azure/azure-resource-manager-rpc/blob/master/v1.0/Addendum.md#provisioningstate-property. + """ + + SUCCEEDED = "Succeeded" + FAILED = "Failed" + CREATING = "Creating" + DELETING = "Deleting" + UPDATING = "Updating" + CANCELED = "Canceled" + + +class ExpansionJobStatusType(str, Enum, metaclass=CaseInsensitiveEnumMeta): + """The operational state of the expansion job. InProgress indicates the expansion is still + running. Completed indicates expansion finished successfully. Failed means the expansion was + unable to complete due to a fatal error. Deleting indicates the expansion is being rolled back. + """ + + IN_PROGRESS = "InProgress" + COMPLETED = "Completed" + FAILED = "Failed" + DELETING = "Deleting" + ROLLING_BACK = "RollingBack" + + class FilesystemSubnetStatusType(str, Enum, metaclass=CaseInsensitiveEnumMeta): """The status of the AML file system subnet check.""" diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/__init__.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/__init__.py index 356469ae8c49..bf2a50f44406 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/__init__.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/__init__.py @@ -17,6 +17,7 @@ from ._auto_export_jobs_operations import AutoExportJobsOperations # type: ignore from ._import_jobs_operations import ImportJobsOperations # type: ignore from ._auto_import_jobs_operations import AutoImportJobsOperations # type: ignore +from ._expansion_jobs_operations import ExpansionJobsOperations # type: ignore from ._storage_cache_management_client_operations import _StorageCacheManagementClientOperationsMixin # type: ignore # pylint: disable=unused-import from ._operations import Operations # type: ignore from ._skus_operations import SkusOperations # type: ignore @@ -36,6 +37,7 @@ "AutoExportJobsOperations", "ImportJobsOperations", "AutoImportJobsOperations", + "ExpansionJobsOperations", "Operations", "SkusOperations", "UsageModelsOperations", diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_aml_filesystems_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_aml_filesystems_operations.py index ff0c02f2b24b..164552124ae4 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_aml_filesystems_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_aml_filesystems_operations.py @@ -35,9 +35,9 @@ from .._configuration import StorageCacheManagementClientConfiguration from .._utils.serialization import Deserializer, Serializer -List = list T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, dict[str, Any]], Any]] +List = list _SERIALIZER = Serializer() _SERIALIZER.client_side_validation = False @@ -47,7 +47,7 @@ def build_list_request(subscription_id: str, **kwargs: Any) -> HttpRequest: _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -71,7 +71,7 @@ def build_list_by_resource_group_request(resource_group_name: str, subscription_ _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -103,7 +103,7 @@ def build_delete_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -143,7 +143,7 @@ def build_get_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -183,7 +183,7 @@ def build_create_or_update_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") @@ -226,7 +226,7 @@ def build_update_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") @@ -269,7 +269,7 @@ def build_archive_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") @@ -312,7 +312,7 @@ def build_cancel_archive_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_asc_operations_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_asc_operations_operations.py index 05c0a2b0e92d..d4a4a8b44255 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_asc_operations_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_asc_operations_operations.py @@ -30,6 +30,7 @@ T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, dict[str, Any]], Any]] +List = list _SERIALIZER = Serializer() _SERIALIZER.client_side_validation = False @@ -39,7 +40,7 @@ def build_get_request(location: str, operation_id: str, subscription_id: str, ** _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_asc_usages_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_asc_usages_operations.py index ab007cd78de5..00f9458f277b 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_asc_usages_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_asc_usages_operations.py @@ -29,9 +29,9 @@ from .._configuration import StorageCacheManagementClientConfiguration from .._utils.serialization import Deserializer, Serializer -List = list T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, dict[str, Any]], Any]] +List = list _SERIALIZER = Serializer() _SERIALIZER.client_side_validation = False @@ -41,7 +41,7 @@ def build_list_request(location: str, subscription_id: str, **kwargs: Any) -> Ht _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_auto_export_jobs_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_auto_export_jobs_operations.py index 4f819f13388c..9d3e61d4c938 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_auto_export_jobs_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_auto_export_jobs_operations.py @@ -37,6 +37,7 @@ T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, dict[str, Any]], Any]] +List = list _SERIALIZER = Serializer() _SERIALIZER.client_side_validation = False @@ -48,7 +49,7 @@ def build_delete_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -96,7 +97,7 @@ def build_get_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -144,7 +145,7 @@ def build_create_or_update_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") @@ -195,7 +196,7 @@ def build_update_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") @@ -246,7 +247,7 @@ def build_list_by_aml_filesystem_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -343,7 +344,10 @@ def _delete_initial( except (StreamConsumedError, StreamClosedError): pass map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) response_headers = {} @@ -475,7 +479,10 @@ def get( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) deserialized = self._deserialize("AutoExportJob", pipeline_response.http_response) @@ -544,7 +551,10 @@ def _create_or_update_initial( except (StreamConsumedError, StreamClosedError): pass map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) response_headers = {} @@ -768,7 +778,10 @@ def _update_initial( except (StreamConsumedError, StreamClosedError): pass map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) response_headers = {} diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_auto_import_jobs_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_auto_import_jobs_operations.py index 0ee114988511..be76de2d00fa 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_auto_import_jobs_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_auto_import_jobs_operations.py @@ -37,6 +37,7 @@ T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, dict[str, Any]], Any]] +List = list _SERIALIZER = Serializer() _SERIALIZER.client_side_validation = False @@ -48,7 +49,7 @@ def build_delete_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -96,7 +97,7 @@ def build_get_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -144,7 +145,7 @@ def build_create_or_update_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") @@ -195,7 +196,7 @@ def build_update_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") @@ -246,7 +247,7 @@ def build_list_by_aml_filesystem_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -343,7 +344,10 @@ def _delete_initial( except (StreamConsumedError, StreamClosedError): pass map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) response_headers = {} @@ -475,7 +479,10 @@ def get( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) deserialized = self._deserialize("AutoImportJob", pipeline_response.http_response) @@ -544,7 +551,10 @@ def _create_or_update_initial( except (StreamConsumedError, StreamClosedError): pass map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) response_headers = {} @@ -768,7 +778,10 @@ def _update_initial( except (StreamConsumedError, StreamClosedError): pass map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) response_headers = {} diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_caches_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_caches_operations.py index 5aceb9fe1b3e..ef781dcb10f3 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_caches_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_caches_operations.py @@ -35,9 +35,9 @@ from .._configuration import StorageCacheManagementClientConfiguration from .._utils.serialization import Deserializer, Serializer -List = list T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, dict[str, Any]], Any]] +List = list _SERIALIZER = Serializer() _SERIALIZER.client_side_validation = False @@ -47,7 +47,7 @@ def build_list_request(subscription_id: str, **kwargs: Any) -> HttpRequest: _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -71,7 +71,7 @@ def build_list_by_resource_group_request(resource_group_name: str, subscription_ _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -101,7 +101,7 @@ def build_delete_request(resource_group_name: str, cache_name: str, subscription _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -132,7 +132,7 @@ def build_get_request(resource_group_name: str, cache_name: str, subscription_id _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -165,7 +165,7 @@ def build_create_or_update_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") @@ -199,7 +199,7 @@ def build_update_request(resource_group_name: str, cache_name: str, subscription _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") @@ -235,7 +235,7 @@ def build_debug_info_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -266,7 +266,7 @@ def build_flush_request(resource_group_name: str, cache_name: str, subscription_ _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -297,7 +297,7 @@ def build_start_request(resource_group_name: str, cache_name: str, subscription_ _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -328,7 +328,7 @@ def build_stop_request(resource_group_name: str, cache_name: str, subscription_i _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -361,7 +361,7 @@ def build_start_priming_job_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") @@ -397,7 +397,7 @@ def build_stop_priming_job_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") @@ -433,7 +433,7 @@ def build_pause_priming_job_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") @@ -469,7 +469,7 @@ def build_resume_priming_job_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") @@ -505,7 +505,7 @@ def build_upgrade_firmware_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -538,7 +538,7 @@ def build_space_allocation_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_expansion_jobs_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_expansion_jobs_operations.py new file mode 100644 index 000000000000..be8a95d17a3b --- /dev/null +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_expansion_jobs_operations.py @@ -0,0 +1,1029 @@ +# pylint: disable=line-too-long,useless-suppression,too-many-lines +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- +from collections.abc import MutableMapping +from io import IOBase +from typing import Any, Callable, IO, Iterator, Optional, TypeVar, Union, cast, overload +import urllib.parse + +from azure.core import PipelineClient +from azure.core.exceptions import ( + ClientAuthenticationError, + HttpResponseError, + ResourceExistsError, + ResourceNotFoundError, + ResourceNotModifiedError, + StreamClosedError, + StreamConsumedError, + map_error, +) +from azure.core.paging import ItemPaged +from azure.core.pipeline import PipelineResponse +from azure.core.polling import LROPoller, NoPolling, PollingMethod +from azure.core.rest import HttpRequest, HttpResponse +from azure.core.tracing.decorator import distributed_trace +from azure.core.utils import case_insensitive_dict +from azure.mgmt.core.exceptions import ARMErrorFormat +from azure.mgmt.core.polling.arm_polling import ARMPolling + +from .. import models as _models +from .._configuration import StorageCacheManagementClientConfiguration +from .._utils.serialization import Deserializer, Serializer + +T = TypeVar("T") +ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, dict[str, Any]], Any]] +List = list + +_SERIALIZER = Serializer() +_SERIALIZER.client_side_validation = False + + +def build_delete_request( + resource_group_name: str, aml_filesystem_name: str, expansion_job_name: str, subscription_id: str, **kwargs: Any +) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = kwargs.pop( + "template_url", + "/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.StorageCache/amlFilesystems/{amlFilesystemName}/expansionJobs/{expansionJobName}", + ) + path_format_arguments = { + "resourceGroupName": _SERIALIZER.url( + "resource_group_name", resource_group_name, "str", max_length=90, min_length=1 + ), + "amlFilesystemName": _SERIALIZER.url( + "aml_filesystem_name", + aml_filesystem_name, + "str", + max_length=80, + min_length=2, + pattern=r"^[0-9a-zA-Z][-0-9a-zA-Z_]{0,78}[0-9a-zA-Z]$", + ), + "subscriptionId": _SERIALIZER.url("subscription_id", subscription_id, "str", min_length=1), + "expansionJobName": _SERIALIZER.url( + "expansion_job_name", + expansion_job_name, + "str", + max_length=80, + min_length=2, + pattern=r"^[0-9a-zA-Z][-0-9a-zA-Z_]{0,78}[0-9a-zA-Z]$", + ), + } + + _url: str = _url.format(**path_format_arguments) # type: ignore + + # Construct parameters + _params["api-version"] = _SERIALIZER.query("api_version", api_version, "str") + + # Construct headers + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest(method="DELETE", url=_url, params=_params, headers=_headers, **kwargs) + + +def build_get_request( + resource_group_name: str, aml_filesystem_name: str, expansion_job_name: str, subscription_id: str, **kwargs: Any +) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = kwargs.pop( + "template_url", + "/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.StorageCache/amlFilesystems/{amlFilesystemName}/expansionJobs/{expansionJobName}", + ) + path_format_arguments = { + "resourceGroupName": _SERIALIZER.url( + "resource_group_name", resource_group_name, "str", max_length=90, min_length=1 + ), + "amlFilesystemName": _SERIALIZER.url( + "aml_filesystem_name", + aml_filesystem_name, + "str", + max_length=80, + min_length=2, + pattern=r"^[0-9a-zA-Z][-0-9a-zA-Z_]{0,78}[0-9a-zA-Z]$", + ), + "subscriptionId": _SERIALIZER.url("subscription_id", subscription_id, "str", min_length=1), + "expansionJobName": _SERIALIZER.url( + "expansion_job_name", + expansion_job_name, + "str", + max_length=80, + min_length=2, + pattern=r"^[0-9a-zA-Z][-0-9a-zA-Z_]{0,78}[0-9a-zA-Z]$", + ), + } + + _url: str = _url.format(**path_format_arguments) # type: ignore + + # Construct parameters + _params["api-version"] = _SERIALIZER.query("api_version", api_version, "str") + + # Construct headers + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest(method="GET", url=_url, params=_params, headers=_headers, **kwargs) + + +def build_create_or_update_request( + resource_group_name: str, aml_filesystem_name: str, expansion_job_name: str, subscription_id: str, **kwargs: Any +) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) + content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = kwargs.pop( + "template_url", + "/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.StorageCache/amlFilesystems/{amlFilesystemName}/expansionJobs/{expansionJobName}", + ) + path_format_arguments = { + "resourceGroupName": _SERIALIZER.url( + "resource_group_name", resource_group_name, "str", max_length=90, min_length=1 + ), + "subscriptionId": _SERIALIZER.url("subscription_id", subscription_id, "str", min_length=1), + "amlFilesystemName": _SERIALIZER.url( + "aml_filesystem_name", + aml_filesystem_name, + "str", + max_length=80, + min_length=2, + pattern=r"^[0-9a-zA-Z][-0-9a-zA-Z_]{0,78}[0-9a-zA-Z]$", + ), + "expansionJobName": _SERIALIZER.url( + "expansion_job_name", + expansion_job_name, + "str", + max_length=80, + min_length=2, + pattern=r"^[0-9a-zA-Z][-0-9a-zA-Z_]{0,78}[0-9a-zA-Z]$", + ), + } + + _url: str = _url.format(**path_format_arguments) # type: ignore + + # Construct parameters + _params["api-version"] = _SERIALIZER.query("api_version", api_version, "str") + + # Construct headers + if content_type is not None: + _headers["Content-Type"] = _SERIALIZER.header("content_type", content_type, "str") + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest(method="PUT", url=_url, params=_params, headers=_headers, **kwargs) + + +def build_update_request( + resource_group_name: str, aml_filesystem_name: str, expansion_job_name: str, subscription_id: str, **kwargs: Any +) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) + content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = kwargs.pop( + "template_url", + "/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.StorageCache/amlFilesystems/{amlFilesystemName}/expansionJobs/{expansionJobName}", + ) + path_format_arguments = { + "resourceGroupName": _SERIALIZER.url( + "resource_group_name", resource_group_name, "str", max_length=90, min_length=1 + ), + "subscriptionId": _SERIALIZER.url("subscription_id", subscription_id, "str", min_length=1), + "amlFilesystemName": _SERIALIZER.url( + "aml_filesystem_name", + aml_filesystem_name, + "str", + max_length=80, + min_length=2, + pattern=r"^[0-9a-zA-Z][-0-9a-zA-Z_]{0,78}[0-9a-zA-Z]$", + ), + "expansionJobName": _SERIALIZER.url( + "expansion_job_name", + expansion_job_name, + "str", + max_length=80, + min_length=2, + pattern=r"^[0-9a-zA-Z][-0-9a-zA-Z_]{0,78}[0-9a-zA-Z]$", + ), + } + + _url: str = _url.format(**path_format_arguments) # type: ignore + + # Construct parameters + _params["api-version"] = _SERIALIZER.query("api_version", api_version, "str") + + # Construct headers + if content_type is not None: + _headers["Content-Type"] = _SERIALIZER.header("content_type", content_type, "str") + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest(method="PATCH", url=_url, params=_params, headers=_headers, **kwargs) + + +def build_list_by_aml_filesystem_request( + resource_group_name: str, aml_filesystem_name: str, subscription_id: str, **kwargs: Any +) -> HttpRequest: + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) + accept = _headers.pop("Accept", "application/json") + + # Construct URL + _url = kwargs.pop( + "template_url", + "/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.StorageCache/amlFilesystems/{amlFilesystemName}/expansionJobs", + ) + path_format_arguments = { + "resourceGroupName": _SERIALIZER.url( + "resource_group_name", resource_group_name, "str", max_length=90, min_length=1 + ), + "subscriptionId": _SERIALIZER.url("subscription_id", subscription_id, "str", min_length=1), + "amlFilesystemName": _SERIALIZER.url( + "aml_filesystem_name", + aml_filesystem_name, + "str", + max_length=80, + min_length=2, + pattern=r"^[0-9a-zA-Z][-0-9a-zA-Z_]{0,78}[0-9a-zA-Z]$", + ), + } + + _url: str = _url.format(**path_format_arguments) # type: ignore + + # Construct parameters + _params["api-version"] = _SERIALIZER.query("api_version", api_version, "str") + + # Construct headers + _headers["Accept"] = _SERIALIZER.header("accept", accept, "str") + + return HttpRequest(method="GET", url=_url, params=_params, headers=_headers, **kwargs) + + +class ExpansionJobsOperations: + """ + .. warning:: + **DO NOT** instantiate this class directly. + + Instead, you should access the following operations through + :class:`~azure.mgmt.storagecache.StorageCacheManagementClient`'s + :attr:`expansion_jobs` attribute. + """ + + models = _models + + def __init__(self, *args, **kwargs) -> None: + input_args = list(args) + self._client: PipelineClient = input_args.pop(0) if input_args else kwargs.pop("client") + self._config: StorageCacheManagementClientConfiguration = ( + input_args.pop(0) if input_args else kwargs.pop("config") + ) + self._serialize: Serializer = input_args.pop(0) if input_args else kwargs.pop("serializer") + self._deserialize: Deserializer = input_args.pop(0) if input_args else kwargs.pop("deserializer") + + def _delete_initial( + self, resource_group_name: str, aml_filesystem_name: str, expansion_job_name: str, **kwargs: Any + ) -> Iterator[bytes]: + error_map: MutableMapping = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._config.api_version)) + cls: ClsType[Iterator[bytes]] = kwargs.pop("cls", None) + + _request = build_delete_request( + resource_group_name=resource_group_name, + aml_filesystem_name=aml_filesystem_name, + expansion_job_name=expansion_job_name, + subscription_id=self._config.subscription_id, + api_version=api_version, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _decompress = kwargs.pop("decompress", True) + _stream = True + pipeline_response: PipelineResponse = self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [202, 204]: + try: + response.read() # Load the body in memory and close the socket + except (StreamConsumedError, StreamClosedError): + pass + map_error(status_code=response.status_code, response=response, error_map=error_map) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) + raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) + + response_headers = {} + if response.status_code == 202: + response_headers["Location"] = self._deserialize("str", response.headers.get("Location")) + response_headers["Azure-AsyncOperation"] = self._deserialize( + "str", response.headers.get("Azure-AsyncOperation") + ) + + deserialized = response.stream_download(self._client._pipeline, decompress=_decompress) + + if cls: + return cls(pipeline_response, deserialized, response_headers) # type: ignore + + return deserialized # type: ignore + + @distributed_trace + def begin_delete( + self, resource_group_name: str, aml_filesystem_name: str, expansion_job_name: str, **kwargs: Any + ) -> LROPoller[None]: + """Schedules an expansion job for deletion. + + :param resource_group_name: The name of the resource group. The name is case insensitive. + Required. + :type resource_group_name: str + :param aml_filesystem_name: Name for the AML file system. Allows alphanumerics, underscores, + and hyphens. Start and end with alphanumeric. Required. + :type aml_filesystem_name: str + :param expansion_job_name: Name for the expansion job. Allows alphanumerics, underscores, and + hyphens. Start and end with alphanumeric. Required. + :type expansion_job_name: str + :return: An instance of LROPoller that returns either None or the result of cls(response) + :rtype: ~azure.core.polling.LROPoller[None] + :raises ~azure.core.exceptions.HttpResponseError: + """ + _headers = kwargs.pop("headers", {}) or {} + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._config.api_version)) + cls: ClsType[None] = kwargs.pop("cls", None) + polling: Union[bool, PollingMethod] = kwargs.pop("polling", True) + lro_delay = kwargs.pop("polling_interval", self._config.polling_interval) + cont_token: Optional[str] = kwargs.pop("continuation_token", None) + if cont_token is None: + raw_result = self._delete_initial( + resource_group_name=resource_group_name, + aml_filesystem_name=aml_filesystem_name, + expansion_job_name=expansion_job_name, + api_version=api_version, + cls=lambda x, y, z: x, + headers=_headers, + params=_params, + **kwargs + ) + raw_result.http_response.read() # type: ignore + kwargs.pop("error_map", None) + + def get_long_running_output(pipeline_response): # pylint: disable=inconsistent-return-statements + if cls: + return cls(pipeline_response, None, {}) # type: ignore + + if polling is True: + polling_method: PollingMethod = cast( + PollingMethod, ARMPolling(lro_delay, lro_options={"final-state-via": "location"}, **kwargs) + ) + elif polling is False: + polling_method = cast(PollingMethod, NoPolling()) + else: + polling_method = polling + if cont_token: + return LROPoller[None].from_continuation_token( + polling_method=polling_method, + continuation_token=cont_token, + client=self._client, + deserialization_callback=get_long_running_output, + ) + return LROPoller[None](self._client, raw_result, get_long_running_output, polling_method) # type: ignore + + @distributed_trace + def get( + self, resource_group_name: str, aml_filesystem_name: str, expansion_job_name: str, **kwargs: Any + ) -> _models.ExpansionJob: + """Returns an expansion job. + + :param resource_group_name: The name of the resource group. The name is case insensitive. + Required. + :type resource_group_name: str + :param aml_filesystem_name: Name for the AML file system. Allows alphanumerics, underscores, + and hyphens. Start and end with alphanumeric. Required. + :type aml_filesystem_name: str + :param expansion_job_name: Name for the expansion job. Allows alphanumerics, underscores, and + hyphens. Start and end with alphanumeric. Required. + :type expansion_job_name: str + :return: ExpansionJob or the result of cls(response) + :rtype: ~azure.mgmt.storagecache.models.ExpansionJob + :raises ~azure.core.exceptions.HttpResponseError: + """ + error_map: MutableMapping = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = kwargs.pop("headers", {}) or {} + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._config.api_version)) + cls: ClsType[_models.ExpansionJob] = kwargs.pop("cls", None) + + _request = build_get_request( + resource_group_name=resource_group_name, + aml_filesystem_name=aml_filesystem_name, + expansion_job_name=expansion_job_name, + subscription_id=self._config.subscription_id, + api_version=api_version, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _stream = False + pipeline_response: PipelineResponse = self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [200]: + map_error(status_code=response.status_code, response=response, error_map=error_map) + raise HttpResponseError(response=response, error_format=ARMErrorFormat) + + deserialized = self._deserialize("ExpansionJob", pipeline_response.http_response) + + if cls: + return cls(pipeline_response, deserialized, {}) # type: ignore + + return deserialized # type: ignore + + def _create_or_update_initial( + self, + resource_group_name: str, + aml_filesystem_name: str, + expansion_job_name: str, + expansion_job: Union[_models.ExpansionJob, IO[bytes]], + **kwargs: Any + ) -> Iterator[bytes]: + error_map: MutableMapping = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._config.api_version)) + content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) + cls: ClsType[Iterator[bytes]] = kwargs.pop("cls", None) + + content_type = content_type or "application/json" + _json = None + _content = None + if isinstance(expansion_job, (IOBase, bytes)): + _content = expansion_job + else: + _json = self._serialize.body(expansion_job, "ExpansionJob") + + _request = build_create_or_update_request( + resource_group_name=resource_group_name, + aml_filesystem_name=aml_filesystem_name, + expansion_job_name=expansion_job_name, + subscription_id=self._config.subscription_id, + api_version=api_version, + content_type=content_type, + json=_json, + content=_content, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _decompress = kwargs.pop("decompress", True) + _stream = True + pipeline_response: PipelineResponse = self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 201]: + try: + response.read() # Load the body in memory and close the socket + except (StreamConsumedError, StreamClosedError): + pass + map_error(status_code=response.status_code, response=response, error_map=error_map) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) + raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) + + response_headers = {} + if response.status_code == 201: + response_headers["Location"] = self._deserialize("str", response.headers.get("Location")) + response_headers["Azure-AsyncOperation"] = self._deserialize( + "str", response.headers.get("Azure-AsyncOperation") + ) + + deserialized = response.stream_download(self._client._pipeline, decompress=_decompress) + + if cls: + return cls(pipeline_response, deserialized, response_headers) # type: ignore + + return deserialized # type: ignore + + @overload + def begin_create_or_update( + self, + resource_group_name: str, + aml_filesystem_name: str, + expansion_job_name: str, + expansion_job: _models.ExpansionJob, + *, + content_type: str = "application/json", + **kwargs: Any + ) -> LROPoller[_models.ExpansionJob]: + """Create or update an expansion job. + + :param resource_group_name: The name of the resource group. The name is case insensitive. + Required. + :type resource_group_name: str + :param aml_filesystem_name: Name for the AML file system. Allows alphanumerics, underscores, + and hyphens. Start and end with alphanumeric. Required. + :type aml_filesystem_name: str + :param expansion_job_name: Name for the expansion job. Allows alphanumerics, underscores, and + hyphens. Start and end with alphanumeric. Required. + :type expansion_job_name: str + :param expansion_job: Object containing the user-selectable properties of the expansion job. If + read-only properties are included, they must match the existing values of those properties. + Required. + :type expansion_job: ~azure.mgmt.storagecache.models.ExpansionJob + :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. + Default value is "application/json". + :paramtype content_type: str + :return: An instance of LROPoller that returns either ExpansionJob or the result of + cls(response) + :rtype: ~azure.core.polling.LROPoller[~azure.mgmt.storagecache.models.ExpansionJob] + :raises ~azure.core.exceptions.HttpResponseError: + """ + + @overload + def begin_create_or_update( + self, + resource_group_name: str, + aml_filesystem_name: str, + expansion_job_name: str, + expansion_job: IO[bytes], + *, + content_type: str = "application/json", + **kwargs: Any + ) -> LROPoller[_models.ExpansionJob]: + """Create or update an expansion job. + + :param resource_group_name: The name of the resource group. The name is case insensitive. + Required. + :type resource_group_name: str + :param aml_filesystem_name: Name for the AML file system. Allows alphanumerics, underscores, + and hyphens. Start and end with alphanumeric. Required. + :type aml_filesystem_name: str + :param expansion_job_name: Name for the expansion job. Allows alphanumerics, underscores, and + hyphens. Start and end with alphanumeric. Required. + :type expansion_job_name: str + :param expansion_job: Object containing the user-selectable properties of the expansion job. If + read-only properties are included, they must match the existing values of those properties. + Required. + :type expansion_job: IO[bytes] + :keyword content_type: Body Parameter content-type. Content type parameter for binary body. + Default value is "application/json". + :paramtype content_type: str + :return: An instance of LROPoller that returns either ExpansionJob or the result of + cls(response) + :rtype: ~azure.core.polling.LROPoller[~azure.mgmt.storagecache.models.ExpansionJob] + :raises ~azure.core.exceptions.HttpResponseError: + """ + + @distributed_trace + def begin_create_or_update( + self, + resource_group_name: str, + aml_filesystem_name: str, + expansion_job_name: str, + expansion_job: Union[_models.ExpansionJob, IO[bytes]], + **kwargs: Any + ) -> LROPoller[_models.ExpansionJob]: + """Create or update an expansion job. + + :param resource_group_name: The name of the resource group. The name is case insensitive. + Required. + :type resource_group_name: str + :param aml_filesystem_name: Name for the AML file system. Allows alphanumerics, underscores, + and hyphens. Start and end with alphanumeric. Required. + :type aml_filesystem_name: str + :param expansion_job_name: Name for the expansion job. Allows alphanumerics, underscores, and + hyphens. Start and end with alphanumeric. Required. + :type expansion_job_name: str + :param expansion_job: Object containing the user-selectable properties of the expansion job. If + read-only properties are included, they must match the existing values of those properties. Is + either a ExpansionJob type or a IO[bytes] type. Required. + :type expansion_job: ~azure.mgmt.storagecache.models.ExpansionJob or IO[bytes] + :return: An instance of LROPoller that returns either ExpansionJob or the result of + cls(response) + :rtype: ~azure.core.polling.LROPoller[~azure.mgmt.storagecache.models.ExpansionJob] + :raises ~azure.core.exceptions.HttpResponseError: + """ + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._config.api_version)) + content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) + cls: ClsType[_models.ExpansionJob] = kwargs.pop("cls", None) + polling: Union[bool, PollingMethod] = kwargs.pop("polling", True) + lro_delay = kwargs.pop("polling_interval", self._config.polling_interval) + cont_token: Optional[str] = kwargs.pop("continuation_token", None) + if cont_token is None: + raw_result = self._create_or_update_initial( + resource_group_name=resource_group_name, + aml_filesystem_name=aml_filesystem_name, + expansion_job_name=expansion_job_name, + expansion_job=expansion_job, + api_version=api_version, + content_type=content_type, + cls=lambda x, y, z: x, + headers=_headers, + params=_params, + **kwargs + ) + raw_result.http_response.read() # type: ignore + kwargs.pop("error_map", None) + + def get_long_running_output(pipeline_response): + deserialized = self._deserialize("ExpansionJob", pipeline_response.http_response) + if cls: + return cls(pipeline_response, deserialized, {}) # type: ignore + return deserialized + + if polling is True: + polling_method: PollingMethod = cast( + PollingMethod, ARMPolling(lro_delay, lro_options={"final-state-via": "azure-async-operation"}, **kwargs) + ) + elif polling is False: + polling_method = cast(PollingMethod, NoPolling()) + else: + polling_method = polling + if cont_token: + return LROPoller[_models.ExpansionJob].from_continuation_token( + polling_method=polling_method, + continuation_token=cont_token, + client=self._client, + deserialization_callback=get_long_running_output, + ) + return LROPoller[_models.ExpansionJob]( + self._client, raw_result, get_long_running_output, polling_method # type: ignore + ) + + def _update_initial( + self, + resource_group_name: str, + aml_filesystem_name: str, + expansion_job_name: str, + expansion_job: Union[_models.ExpansionJobUpdate, IO[bytes]], + **kwargs: Any + ) -> Iterator[bytes]: + error_map: MutableMapping = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._config.api_version)) + content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) + cls: ClsType[Iterator[bytes]] = kwargs.pop("cls", None) + + content_type = content_type or "application/json" + _json = None + _content = None + if isinstance(expansion_job, (IOBase, bytes)): + _content = expansion_job + else: + _json = self._serialize.body(expansion_job, "ExpansionJobUpdate") + + _request = build_update_request( + resource_group_name=resource_group_name, + aml_filesystem_name=aml_filesystem_name, + expansion_job_name=expansion_job_name, + subscription_id=self._config.subscription_id, + api_version=api_version, + content_type=content_type, + json=_json, + content=_content, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + _decompress = kwargs.pop("decompress", True) + _stream = True + pipeline_response: PipelineResponse = self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + + response = pipeline_response.http_response + + if response.status_code not in [200, 202]: + try: + response.read() # Load the body in memory and close the socket + except (StreamConsumedError, StreamClosedError): + pass + map_error(status_code=response.status_code, response=response, error_map=error_map) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) + raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) + + response_headers = {} + if response.status_code == 202: + response_headers["Location"] = self._deserialize("str", response.headers.get("Location")) + response_headers["Azure-AsyncOperation"] = self._deserialize( + "str", response.headers.get("Azure-AsyncOperation") + ) + + deserialized = response.stream_download(self._client._pipeline, decompress=_decompress) + + if cls: + return cls(pipeline_response, deserialized, response_headers) # type: ignore + + return deserialized # type: ignore + + @overload + def begin_update( + self, + resource_group_name: str, + aml_filesystem_name: str, + expansion_job_name: str, + expansion_job: _models.ExpansionJobUpdate, + *, + content_type: str = "application/json", + **kwargs: Any + ) -> LROPoller[_models.ExpansionJob]: + """Update an expansion job instance. + + :param resource_group_name: The name of the resource group. The name is case insensitive. + Required. + :type resource_group_name: str + :param aml_filesystem_name: Name for the AML file system. Allows alphanumerics, underscores, + and hyphens. Start and end with alphanumeric. Required. + :type aml_filesystem_name: str + :param expansion_job_name: Name for the expansion job. Allows alphanumerics, underscores, and + hyphens. Start and end with alphanumeric. Required. + :type expansion_job_name: str + :param expansion_job: Object containing the user-selectable properties of the expansion job. If + read-only properties are included, they must match the existing values of those properties. + Required. + :type expansion_job: ~azure.mgmt.storagecache.models.ExpansionJobUpdate + :keyword content_type: Body Parameter content-type. Content type parameter for JSON body. + Default value is "application/json". + :paramtype content_type: str + :return: An instance of LROPoller that returns either ExpansionJob or the result of + cls(response) + :rtype: ~azure.core.polling.LROPoller[~azure.mgmt.storagecache.models.ExpansionJob] + :raises ~azure.core.exceptions.HttpResponseError: + """ + + @overload + def begin_update( + self, + resource_group_name: str, + aml_filesystem_name: str, + expansion_job_name: str, + expansion_job: IO[bytes], + *, + content_type: str = "application/json", + **kwargs: Any + ) -> LROPoller[_models.ExpansionJob]: + """Update an expansion job instance. + + :param resource_group_name: The name of the resource group. The name is case insensitive. + Required. + :type resource_group_name: str + :param aml_filesystem_name: Name for the AML file system. Allows alphanumerics, underscores, + and hyphens. Start and end with alphanumeric. Required. + :type aml_filesystem_name: str + :param expansion_job_name: Name for the expansion job. Allows alphanumerics, underscores, and + hyphens. Start and end with alphanumeric. Required. + :type expansion_job_name: str + :param expansion_job: Object containing the user-selectable properties of the expansion job. If + read-only properties are included, they must match the existing values of those properties. + Required. + :type expansion_job: IO[bytes] + :keyword content_type: Body Parameter content-type. Content type parameter for binary body. + Default value is "application/json". + :paramtype content_type: str + :return: An instance of LROPoller that returns either ExpansionJob or the result of + cls(response) + :rtype: ~azure.core.polling.LROPoller[~azure.mgmt.storagecache.models.ExpansionJob] + :raises ~azure.core.exceptions.HttpResponseError: + """ + + @distributed_trace + def begin_update( + self, + resource_group_name: str, + aml_filesystem_name: str, + expansion_job_name: str, + expansion_job: Union[_models.ExpansionJobUpdate, IO[bytes]], + **kwargs: Any + ) -> LROPoller[_models.ExpansionJob]: + """Update an expansion job instance. + + :param resource_group_name: The name of the resource group. The name is case insensitive. + Required. + :type resource_group_name: str + :param aml_filesystem_name: Name for the AML file system. Allows alphanumerics, underscores, + and hyphens. Start and end with alphanumeric. Required. + :type aml_filesystem_name: str + :param expansion_job_name: Name for the expansion job. Allows alphanumerics, underscores, and + hyphens. Start and end with alphanumeric. Required. + :type expansion_job_name: str + :param expansion_job: Object containing the user-selectable properties of the expansion job. If + read-only properties are included, they must match the existing values of those properties. Is + either a ExpansionJobUpdate type or a IO[bytes] type. Required. + :type expansion_job: ~azure.mgmt.storagecache.models.ExpansionJobUpdate or IO[bytes] + :return: An instance of LROPoller that returns either ExpansionJob or the result of + cls(response) + :rtype: ~azure.core.polling.LROPoller[~azure.mgmt.storagecache.models.ExpansionJob] + :raises ~azure.core.exceptions.HttpResponseError: + """ + _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._config.api_version)) + content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) + cls: ClsType[_models.ExpansionJob] = kwargs.pop("cls", None) + polling: Union[bool, PollingMethod] = kwargs.pop("polling", True) + lro_delay = kwargs.pop("polling_interval", self._config.polling_interval) + cont_token: Optional[str] = kwargs.pop("continuation_token", None) + if cont_token is None: + raw_result = self._update_initial( + resource_group_name=resource_group_name, + aml_filesystem_name=aml_filesystem_name, + expansion_job_name=expansion_job_name, + expansion_job=expansion_job, + api_version=api_version, + content_type=content_type, + cls=lambda x, y, z: x, + headers=_headers, + params=_params, + **kwargs + ) + raw_result.http_response.read() # type: ignore + kwargs.pop("error_map", None) + + def get_long_running_output(pipeline_response): + deserialized = self._deserialize("ExpansionJob", pipeline_response.http_response) + if cls: + return cls(pipeline_response, deserialized, {}) # type: ignore + return deserialized + + if polling is True: + polling_method: PollingMethod = cast( + PollingMethod, ARMPolling(lro_delay, lro_options={"final-state-via": "azure-async-operation"}, **kwargs) + ) + elif polling is False: + polling_method = cast(PollingMethod, NoPolling()) + else: + polling_method = polling + if cont_token: + return LROPoller[_models.ExpansionJob].from_continuation_token( + polling_method=polling_method, + continuation_token=cont_token, + client=self._client, + deserialization_callback=get_long_running_output, + ) + return LROPoller[_models.ExpansionJob]( + self._client, raw_result, get_long_running_output, polling_method # type: ignore + ) + + @distributed_trace + def list_by_aml_filesystem( + self, resource_group_name: str, aml_filesystem_name: str, **kwargs: Any + ) -> ItemPaged["_models.ExpansionJob"]: + """Returns all the expansion jobs the user has access to under an AML File System. + + :param resource_group_name: The name of the resource group. The name is case insensitive. + Required. + :type resource_group_name: str + :param aml_filesystem_name: Name for the AML file system. Allows alphanumerics, underscores, + and hyphens. Start and end with alphanumeric. Required. + :type aml_filesystem_name: str + :return: An iterator like instance of either ExpansionJob or the result of cls(response) + :rtype: ~azure.core.paging.ItemPaged[~azure.mgmt.storagecache.models.ExpansionJob] + :raises ~azure.core.exceptions.HttpResponseError: + """ + _headers = kwargs.pop("headers", {}) or {} + _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) + + api_version: str = kwargs.pop("api_version", _params.pop("api-version", self._config.api_version)) + cls: ClsType[_models.ExpansionJobsListResult] = kwargs.pop("cls", None) + + error_map: MutableMapping = { + 401: ClientAuthenticationError, + 404: ResourceNotFoundError, + 409: ResourceExistsError, + 304: ResourceNotModifiedError, + } + error_map.update(kwargs.pop("error_map", {}) or {}) + + def prepare_request(next_link=None): + if not next_link: + + _request = build_list_by_aml_filesystem_request( + resource_group_name=resource_group_name, + aml_filesystem_name=aml_filesystem_name, + subscription_id=self._config.subscription_id, + api_version=api_version, + headers=_headers, + params=_params, + ) + _request.url = self._client.format_url(_request.url) + + else: + # make call to next link with the client's api-version + _parsed_next_link = urllib.parse.urlparse(next_link) + _next_request_params = case_insensitive_dict( + { + key: [urllib.parse.quote(v) for v in value] + for key, value in urllib.parse.parse_qs(_parsed_next_link.query).items() + } + ) + _next_request_params["api-version"] = self._config.api_version + _request = HttpRequest( + "GET", urllib.parse.urljoin(next_link, _parsed_next_link.path), params=_next_request_params + ) + _request.url = self._client.format_url(_request.url) + _request.method = "GET" + return _request + + def extract_data(pipeline_response): + deserialized = self._deserialize("ExpansionJobsListResult", pipeline_response) + list_of_elem = deserialized.value + if cls: + list_of_elem = cls(list_of_elem) # type: ignore + return deserialized.next_link or None, iter(list_of_elem) + + def get_next(next_link=None): + _request = prepare_request(next_link) + + _stream = False + pipeline_response: PipelineResponse = self._client._pipeline.run( # pylint: disable=protected-access + _request, stream=_stream, **kwargs + ) + response = pipeline_response.http_response + + if response.status_code not in [200]: + map_error(status_code=response.status_code, response=response, error_map=error_map) + raise HttpResponseError(response=response, error_format=ARMErrorFormat) + + return pipeline_response + + return ItemPaged(get_next, extract_data) diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_import_jobs_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_import_jobs_operations.py index 78b0ff93c369..93af9f175c6e 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_import_jobs_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_import_jobs_operations.py @@ -37,6 +37,7 @@ T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, dict[str, Any]], Any]] +List = list _SERIALIZER = Serializer() _SERIALIZER.client_side_validation = False @@ -48,7 +49,7 @@ def build_delete_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -96,7 +97,7 @@ def build_get_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -144,7 +145,7 @@ def build_create_or_update_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") @@ -195,7 +196,7 @@ def build_update_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") @@ -246,7 +247,7 @@ def build_list_by_aml_filesystem_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -343,7 +344,10 @@ def _delete_initial( except (StreamConsumedError, StreamClosedError): pass map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) response_headers = {} @@ -475,7 +479,10 @@ def get( if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) deserialized = self._deserialize("ImportJob", pipeline_response.http_response) @@ -544,7 +551,10 @@ def _create_or_update_initial( except (StreamConsumedError, StreamClosedError): pass map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) response_headers = {} @@ -765,7 +775,10 @@ def _update_initial( except (StreamConsumedError, StreamClosedError): pass map_error(status_code=response.status_code, response=response, error_map=error_map) - error = self._deserialize.failsafe_deserialize(_models.ErrorResponse, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.ErrorResponse, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) response_headers = {} diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_operations.py index ba6907cd15c4..69c3aa5318db 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_operations.py @@ -29,9 +29,9 @@ from .._configuration import StorageCacheManagementClientConfiguration from .._utils.serialization import Deserializer, Serializer -List = list T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, dict[str, Any]], Any]] +List = list _SERIALIZER = Serializer() _SERIALIZER.client_side_validation = False @@ -41,7 +41,7 @@ def build_list_request(**kwargs: Any) -> HttpRequest: _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_skus_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_skus_operations.py index bd15f0a740f7..082e64d0d21b 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_skus_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_skus_operations.py @@ -29,9 +29,9 @@ from .._configuration import StorageCacheManagementClientConfiguration from .._utils.serialization import Deserializer, Serializer -List = list T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, dict[str, Any]], Any]] +List = list _SERIALIZER = Serializer() _SERIALIZER.client_side_validation = False @@ -41,7 +41,7 @@ def build_list_request(subscription_id: str, **kwargs: Any) -> HttpRequest: _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_storage_cache_management_client_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_storage_cache_management_client_operations.py index 9a2b011d2178..6e22883e97e5 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_storage_cache_management_client_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_storage_cache_management_client_operations.py @@ -31,6 +31,7 @@ T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, dict[str, Any]], Any]] +List = list _SERIALIZER = Serializer() _SERIALIZER.client_side_validation = False @@ -40,7 +41,7 @@ def build_check_aml_fs_subnets_request(subscription_id: str, **kwargs: Any) -> H _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") @@ -71,7 +72,7 @@ def build_get_required_aml_fs_subnets_size_request( # pylint: disable=name-too- _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") @@ -207,7 +208,10 @@ def check_aml_fs_subnets( # pylint: disable=inconsistent-return-statements map_error(status_code=response.status_code, response=response, error_map=error_map) error = None if response.status_code == 400: - error = self._deserialize.failsafe_deserialize(_models.AmlFilesystemCheckSubnetError, pipeline_response) + error = self._deserialize.failsafe_deserialize( + _models.AmlFilesystemCheckSubnetError, + pipeline_response, + ) raise HttpResponseError(response=response, model=error, error_format=ARMErrorFormat) if cls: diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_storage_target_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_storage_target_operations.py index de7469dae482..bfc09a746ef3 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_storage_target_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_storage_target_operations.py @@ -34,6 +34,7 @@ T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, dict[str, Any]], Any]] +List = list _SERIALIZER = Serializer() _SERIALIZER.client_side_validation = False @@ -45,7 +46,7 @@ def build_flush_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -81,7 +82,7 @@ def build_suspend_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -117,7 +118,7 @@ def build_resume_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -153,7 +154,7 @@ def build_invalidate_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_storage_targets_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_storage_targets_operations.py index 58f067ef70bb..5aa85bd9e475 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_storage_targets_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_storage_targets_operations.py @@ -37,6 +37,7 @@ T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, dict[str, Any]], Any]] +List = list _SERIALIZER = Serializer() _SERIALIZER.client_side_validation = False @@ -48,7 +49,7 @@ def build_dns_refresh_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -84,7 +85,7 @@ def build_list_by_cache_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -123,7 +124,7 @@ def build_delete_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -161,7 +162,7 @@ def build_get_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL @@ -197,7 +198,7 @@ def build_create_or_update_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) content_type: Optional[str] = kwargs.pop("content_type", _headers.pop("Content-Type", None)) accept = _headers.pop("Accept", "application/json") @@ -236,7 +237,7 @@ def build_restore_defaults_request( _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL diff --git a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_usage_models_operations.py b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_usage_models_operations.py index cbc912772030..a5476fd7a18d 100644 --- a/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_usage_models_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/azure/mgmt/storagecache/operations/_usage_models_operations.py @@ -29,9 +29,9 @@ from .._configuration import StorageCacheManagementClientConfiguration from .._utils.serialization import Deserializer, Serializer -List = list T = TypeVar("T") ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, dict[str, Any]], Any]] +List = list _SERIALIZER = Serializer() _SERIALIZER.client_side_validation = False @@ -41,7 +41,7 @@ def build_list_request(subscription_id: str, **kwargs: Any) -> HttpRequest: _headers = case_insensitive_dict(kwargs.pop("headers", {}) or {}) _params = case_insensitive_dict(kwargs.pop("params", {}) or {}) - api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2025-07-01")) + api_version: str = kwargs.pop("api_version", _params.pop("api-version", "2026-01-01")) accept = _headers.pop("Accept", "application/json") # Construct URL diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_archive.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_archive.py index 87a5d0629bdc..3150cbc26a27 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_archive.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_archive.py @@ -37,6 +37,6 @@ def main(): ) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/amlFilesystems_Archive.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/amlFilesystems_Archive.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_cancel_archive.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_cancel_archive.py index f22371c07c4b..33cfa6d2d568 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_cancel_archive.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_cancel_archive.py @@ -37,6 +37,6 @@ def main(): ) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/amlFilesystems_CancelArchive.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/amlFilesystems_CancelArchive.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_create_or_update.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_create_or_update.py index 2a31758c9be2..828cabeaec0e 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_create_or_update.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_create_or_update.py @@ -76,6 +76,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/amlFilesystems_CreateOrUpdate.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/amlFilesystems_CreateOrUpdate.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_delete.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_delete.py index f1ef2430713a..9862b4068c75 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_delete.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_delete.py @@ -37,6 +37,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/amlFilesystems_Delete.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/amlFilesystems_Delete.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_get.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_get.py index 355df4053e90..b603338fb6c1 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_get.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_get.py @@ -38,6 +38,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/amlFilesystems_Get.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/amlFilesystems_Get.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_list.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_list.py index 7747634cf374..39e0edf64706 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_list.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_list.py @@ -36,6 +36,6 @@ def main(): print(item) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/amlFilesystems_List.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/amlFilesystems_List.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_list_by_resource_group.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_list_by_resource_group.py index 648f740af67f..3a49a72a8aa3 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_list_by_resource_group.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_list_by_resource_group.py @@ -38,6 +38,6 @@ def main(): print(item) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/amlFilesystems_ListByResourceGroup.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/amlFilesystems_ListByResourceGroup.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_update.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_update.py index 51a9c4a77ee4..ba770de10db3 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_update.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/aml_filesystems_update.py @@ -58,6 +58,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/amlFilesystems_Update.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/amlFilesystems_Update.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/asc_operations_get.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/asc_operations_get.py index a1583e82eec2..8eefcf6a0b73 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/asc_operations_get.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/asc_operations_get.py @@ -38,6 +38,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/AscOperations_Get.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/AscOperations_Get.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/asc_resource_usages_get.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/asc_resource_usages_get.py index 5d463a2d8b5b..3db12a338374 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/asc_resource_usages_get.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/asc_resource_usages_get.py @@ -38,6 +38,6 @@ def main(): print(item) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/AscResourceUsages_Get.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/AscResourceUsages_Get.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_create_or_update.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_create_or_update.py index 9e89927e4c0a..218093d4b0df 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_create_or_update.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_create_or_update.py @@ -44,6 +44,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/autoExportJobs_CreateOrUpdate.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/autoExportJobs_CreateOrUpdate.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_delete.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_delete.py index f919c5fff544..d1d71c00e214 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_delete.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_delete.py @@ -38,6 +38,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/autoExportJobs_Delete.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/autoExportJobs_Delete.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_get.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_get.py index 257981ffb86b..299b33828d25 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_get.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_get.py @@ -39,6 +39,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/autoExportJobs_Get.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/autoExportJobs_Get.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_list_by_aml_filesystem.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_list_by_aml_filesystem.py index 1ce80b6732bb..fd145cf98ae9 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_list_by_aml_filesystem.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_list_by_aml_filesystem.py @@ -39,6 +39,6 @@ def main(): print(item) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/autoExportJobs_ListByAmlFilesystem.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/autoExportJobs_ListByAmlFilesystem.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_update.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_update.py index 4cde6301270d..0c736e4225a3 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_update.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_export_jobs_update.py @@ -40,6 +40,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/autoExportJobs_Update.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/autoExportJobs_Update.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_create_or_update.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_create_or_update.py index 27ae3bb5770b..1926e1f14f7c 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_create_or_update.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_create_or_update.py @@ -50,6 +50,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/autoImportJobs_CreateOrUpdate.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/autoImportJobs_CreateOrUpdate.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_delete.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_delete.py index 2a593f74472b..c454cf472be8 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_delete.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_delete.py @@ -38,6 +38,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/autoImportJobs_Delete.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/autoImportJobs_Delete.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_get.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_get.py index 3a6733ee9721..9a3921c399cb 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_get.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_get.py @@ -39,6 +39,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/autoImportJobs_Get.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/autoImportJobs_Get.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_list_by_aml_filesystem.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_list_by_aml_filesystem.py index 3a76d7a85057..b13efbd739a4 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_list_by_aml_filesystem.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_list_by_aml_filesystem.py @@ -39,6 +39,6 @@ def main(): print(item) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/autoImportJobs_ListByAmlFilesystem.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/autoImportJobs_ListByAmlFilesystem.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_update.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_update.py index 967845bdbfeb..d86c6afb07cd 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_update.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/auto_import_jobs_update.py @@ -40,6 +40,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/autoImportJobs_Update.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/autoImportJobs_Update.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_create_or_update.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_create_or_update.py index 3fec73231328..72fd0ecbdb42 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_create_or_update.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_create_or_update.py @@ -98,6 +98,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/Caches_CreateOrUpdate.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/Caches_CreateOrUpdate.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_create_or_update_ldap_only.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_create_or_update_ldap_only.py index 34004e1cafe5..87a8f8590783 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_create_or_update_ldap_only.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_create_or_update_ldap_only.py @@ -84,6 +84,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/Caches_CreateOrUpdate_ldap_only.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/Caches_CreateOrUpdate_ldap_only.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_debug_info.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_debug_info.py index b938f804dd19..123f28fcba8d 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_debug_info.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_debug_info.py @@ -37,6 +37,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/Caches_DebugInfo.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/Caches_DebugInfo.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_delete.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_delete.py index 1fb293a5167d..58d2f3c7527e 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_delete.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_delete.py @@ -37,6 +37,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/Caches_Delete.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/Caches_Delete.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_flush.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_flush.py index 55fe806a3cea..bf2194f446fb 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_flush.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_flush.py @@ -37,6 +37,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/Caches_Flush.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/Caches_Flush.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_get.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_get.py index 4ed9e9ab8a4d..8d19b680ab15 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_get.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_get.py @@ -38,6 +38,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/Caches_Get.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/Caches_Get.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_list.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_list.py index 5966873ea14a..544a5ae17263 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_list.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_list.py @@ -36,6 +36,6 @@ def main(): print(item) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/Caches_List.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/Caches_List.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_list_by_resource_group.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_list_by_resource_group.py index 144dfcb40057..e3cdf5d18ea8 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_list_by_resource_group.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_list_by_resource_group.py @@ -38,6 +38,6 @@ def main(): print(item) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/Caches_ListByResourceGroup.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/Caches_ListByResourceGroup.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_start.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_start.py index 654e9f9383b9..03870d936953 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_start.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_start.py @@ -37,6 +37,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/Caches_Start.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/Caches_Start.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_stop.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_stop.py index d0aa95ed8a75..ec5c651c6ce0 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_stop.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_stop.py @@ -37,6 +37,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/Caches_Stop.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/Caches_Stop.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_update.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_update.py index 032d199b149d..233fadaf406d 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_update.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_update.py @@ -38,6 +38,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/Caches_Update.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/Caches_Update.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_update_ldap_only.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_update_ldap_only.py index 1754cd4c09fa..e4b419e90851 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_update_ldap_only.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_update_ldap_only.py @@ -38,6 +38,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/Caches_Update_ldap_only.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/Caches_Update_ldap_only.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_upgrade_firmware.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_upgrade_firmware.py index c7ef98e72a62..1b4e71b9565f 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_upgrade_firmware.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/caches_upgrade_firmware.py @@ -37,6 +37,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/Caches_UpgradeFirmware.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/Caches_UpgradeFirmware.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/check_aml_fs_subnets.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/check_aml_fs_subnets.py index 7e94e154fb4a..86ce75828c7b 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/check_aml_fs_subnets.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/check_aml_fs_subnets.py @@ -34,6 +34,6 @@ def main(): client.check_aml_fs_subnets() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/checkAmlFSSubnets.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/checkAmlFSSubnets.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_create_or_update.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_create_or_update.py new file mode 100644 index 000000000000..1820f622f9b2 --- /dev/null +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_create_or_update.py @@ -0,0 +1,49 @@ +# pylint: disable=line-too-long,useless-suppression +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from azure.identity import DefaultAzureCredential + +from azure.mgmt.storagecache import StorageCacheManagementClient + +""" +# PREREQUISITES + pip install azure-identity + pip install azure-mgmt-storagecache +# USAGE + python expansion_jobs_create_or_update.py + + Before run the sample, please set the values of the client ID, tenant ID and client secret + of the AAD application as environment variables: AZURE_CLIENT_ID, AZURE_TENANT_ID, + AZURE_CLIENT_SECRET. For more info about how to get the value, please see: + https://docs.microsoft.com/azure/active-directory/develop/howto-create-service-principal-portal +""" + + +def main(): + client = StorageCacheManagementClient( + credential=DefaultAzureCredential(), + subscription_id="00000000-0000-0000-0000-000000000000", + ) + + response = client.expansion_jobs.begin_create_or_update( + resource_group_name="scgroup", + aml_filesystem_name="fs1", + expansion_job_name="expansionjob1", + expansion_job={ + "location": "eastus", + "properties": {"newStorageCapacityTiB": 16}, + "tags": {"Dept": "ContosoAds"}, + }, + ).result() + print(response) + + +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/expansionJobs_CreateOrUpdate.json +if __name__ == "__main__": + main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_delete.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_delete.py new file mode 100644 index 000000000000..53953d979f40 --- /dev/null +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_delete.py @@ -0,0 +1,43 @@ +# pylint: disable=line-too-long,useless-suppression +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from azure.identity import DefaultAzureCredential + +from azure.mgmt.storagecache import StorageCacheManagementClient + +""" +# PREREQUISITES + pip install azure-identity + pip install azure-mgmt-storagecache +# USAGE + python expansion_jobs_delete.py + + Before run the sample, please set the values of the client ID, tenant ID and client secret + of the AAD application as environment variables: AZURE_CLIENT_ID, AZURE_TENANT_ID, + AZURE_CLIENT_SECRET. For more info about how to get the value, please see: + https://docs.microsoft.com/azure/active-directory/develop/howto-create-service-principal-portal +""" + + +def main(): + client = StorageCacheManagementClient( + credential=DefaultAzureCredential(), + subscription_id="00000000-0000-0000-0000-000000000000", + ) + + client.expansion_jobs.begin_delete( + resource_group_name="scgroup", + aml_filesystem_name="fs1", + expansion_job_name="expansionjob1", + ).result() + + +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/expansionJobs_Delete.json +if __name__ == "__main__": + main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_get.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_get.py new file mode 100644 index 000000000000..752a86a6c758 --- /dev/null +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_get.py @@ -0,0 +1,44 @@ +# pylint: disable=line-too-long,useless-suppression +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from azure.identity import DefaultAzureCredential + +from azure.mgmt.storagecache import StorageCacheManagementClient + +""" +# PREREQUISITES + pip install azure-identity + pip install azure-mgmt-storagecache +# USAGE + python expansion_jobs_get.py + + Before run the sample, please set the values of the client ID, tenant ID and client secret + of the AAD application as environment variables: AZURE_CLIENT_ID, AZURE_TENANT_ID, + AZURE_CLIENT_SECRET. For more info about how to get the value, please see: + https://docs.microsoft.com/azure/active-directory/develop/howto-create-service-principal-portal +""" + + +def main(): + client = StorageCacheManagementClient( + credential=DefaultAzureCredential(), + subscription_id="00000000-0000-0000-0000-000000000000", + ) + + response = client.expansion_jobs.get( + resource_group_name="scgroup", + aml_filesystem_name="fs1", + expansion_job_name="expansionjob1", + ) + print(response) + + +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/expansionJobs_Get.json +if __name__ == "__main__": + main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_list_by_aml_filesystem.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_list_by_aml_filesystem.py new file mode 100644 index 000000000000..8e189eb99585 --- /dev/null +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_list_by_aml_filesystem.py @@ -0,0 +1,44 @@ +# pylint: disable=line-too-long,useless-suppression +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from azure.identity import DefaultAzureCredential + +from azure.mgmt.storagecache import StorageCacheManagementClient + +""" +# PREREQUISITES + pip install azure-identity + pip install azure-mgmt-storagecache +# USAGE + python expansion_jobs_list_by_aml_filesystem.py + + Before run the sample, please set the values of the client ID, tenant ID and client secret + of the AAD application as environment variables: AZURE_CLIENT_ID, AZURE_TENANT_ID, + AZURE_CLIENT_SECRET. For more info about how to get the value, please see: + https://docs.microsoft.com/azure/active-directory/develop/howto-create-service-principal-portal +""" + + +def main(): + client = StorageCacheManagementClient( + credential=DefaultAzureCredential(), + subscription_id="00000000-0000-0000-0000-000000000000", + ) + + response = client.expansion_jobs.list_by_aml_filesystem( + resource_group_name="scgroup", + aml_filesystem_name="fs1", + ) + for item in response: + print(item) + + +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/expansionJobs_ListByAmlFilesystem.json +if __name__ == "__main__": + main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_update.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_update.py new file mode 100644 index 000000000000..0c2cc74bc6a1 --- /dev/null +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/expansion_jobs_update.py @@ -0,0 +1,45 @@ +# pylint: disable=line-too-long,useless-suppression +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- + +from azure.identity import DefaultAzureCredential + +from azure.mgmt.storagecache import StorageCacheManagementClient + +""" +# PREREQUISITES + pip install azure-identity + pip install azure-mgmt-storagecache +# USAGE + python expansion_jobs_update.py + + Before run the sample, please set the values of the client ID, tenant ID and client secret + of the AAD application as environment variables: AZURE_CLIENT_ID, AZURE_TENANT_ID, + AZURE_CLIENT_SECRET. For more info about how to get the value, please see: + https://docs.microsoft.com/azure/active-directory/develop/howto-create-service-principal-portal +""" + + +def main(): + client = StorageCacheManagementClient( + credential=DefaultAzureCredential(), + subscription_id="00000000-0000-0000-0000-000000000000", + ) + + response = client.expansion_jobs.begin_update( + resource_group_name="scgroup", + aml_filesystem_name="fs1", + expansion_job_name="expansionjob1", + expansion_job={"tags": {"Dept": "ContosoFinance"}}, + ).result() + print(response) + + +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/expansionJobs_Update.json +if __name__ == "__main__": + main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/get_required_aml_fs_subnets_size.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/get_required_aml_fs_subnets_size.py index 16cc90348ece..9bc2412fa406 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/get_required_aml_fs_subnets_size.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/get_required_aml_fs_subnets_size.py @@ -35,6 +35,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/getRequiredAmlFSSubnetsSize.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/getRequiredAmlFSSubnetsSize.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/import_job_update.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/import_job_update.py index 9a8a2089cd12..d3ae45a101a7 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/import_job_update.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/import_job_update.py @@ -40,6 +40,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/importJob_Update.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/importJob_Update.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/import_jobs_create_or_update.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/import_jobs_create_or_update.py index cc4e67f6a62f..b7a2a0a2dd45 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/import_jobs_create_or_update.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/import_jobs_create_or_update.py @@ -44,6 +44,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/importJobs_CreateOrUpdate.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/importJobs_CreateOrUpdate.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/import_jobs_delete.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/import_jobs_delete.py index 9750186359c9..0f25fd0ee74d 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/import_jobs_delete.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/import_jobs_delete.py @@ -38,6 +38,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/importJobs_Delete.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/importJobs_Delete.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/import_jobs_get.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/import_jobs_get.py index 1efa903250a3..875bd89e1fed 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/import_jobs_get.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/import_jobs_get.py @@ -39,6 +39,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/importJobs_Get.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/importJobs_Get.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/import_jobs_list_by_aml_filesystem.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/import_jobs_list_by_aml_filesystem.py index 03409073a56c..be53a94bf109 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/import_jobs_list_by_aml_filesystem.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/import_jobs_list_by_aml_filesystem.py @@ -39,6 +39,6 @@ def main(): print(item) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/importJobs_ListByAmlFilesystem.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/importJobs_ListByAmlFilesystem.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/operations_list.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/operations_list.py index 346ec59916f2..fff4693f7f8e 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/operations_list.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/operations_list.py @@ -36,6 +36,6 @@ def main(): print(item) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/Operations_List.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/Operations_List.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/pause_priming_job.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/pause_priming_job.py index 816683994046..7ca0941ad9af 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/pause_priming_job.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/pause_priming_job.py @@ -37,6 +37,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/PausePrimingJob.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/PausePrimingJob.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/resume_priming_job.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/resume_priming_job.py index 59f3b63f74c3..c1e32b8b3e87 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/resume_priming_job.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/resume_priming_job.py @@ -37,6 +37,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/ResumePrimingJob.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/ResumePrimingJob.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/skus_list.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/skus_list.py index c8d7dd632f4d..bfcdc6d99ac4 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/skus_list.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/skus_list.py @@ -36,6 +36,6 @@ def main(): print(item) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/Skus_List.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/Skus_List.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/space_allocation_post.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/space_allocation_post.py index 73c3f3b7df3f..5db6d22f7123 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/space_allocation_post.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/space_allocation_post.py @@ -37,6 +37,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/SpaceAllocation_Post.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/SpaceAllocation_Post.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/start_priming_job.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/start_priming_job.py index 4729036d0f9f..cc43b7aea412 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/start_priming_job.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/start_priming_job.py @@ -37,6 +37,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/StartPrimingJob.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/StartPrimingJob.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/stop_priming_job.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/stop_priming_job.py index 9f74071eb731..b7891c4ce43f 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/stop_priming_job.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/stop_priming_job.py @@ -37,6 +37,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/StopPrimingJob.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/StopPrimingJob.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_create_or_update.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_create_or_update.py index 3974b23ac0ba..acde55ea3d87 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_create_or_update.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_create_or_update.py @@ -59,6 +59,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/StorageTargets_CreateOrUpdate.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/StorageTargets_CreateOrUpdate.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_create_or_update_blob_nfs.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_create_or_update_blob_nfs.py index cb986314755b..9d0282148ddc 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_create_or_update_blob_nfs.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_create_or_update_blob_nfs.py @@ -51,6 +51,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/StorageTargets_CreateOrUpdate_BlobNfs.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/StorageTargets_CreateOrUpdate_BlobNfs.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_create_or_update_no_junctions.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_create_or_update_no_junctions.py index 88fb2717d3e3..aa4d557f96fc 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_create_or_update_no_junctions.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_create_or_update_no_junctions.py @@ -45,6 +45,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/StorageTargets_CreateOrUpdate_NoJunctions.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/StorageTargets_CreateOrUpdate_NoJunctions.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_delete.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_delete.py index a9eec13d28ec..1916b9242d15 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_delete.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_delete.py @@ -38,6 +38,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/StorageTargets_Delete.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/StorageTargets_Delete.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_dns_refresh.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_dns_refresh.py index 92e335b4cac0..b1499dd79ebd 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_dns_refresh.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_dns_refresh.py @@ -38,6 +38,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/StorageTargets_DnsRefresh.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/StorageTargets_DnsRefresh.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_flush.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_flush.py index 4afa217e2529..248337c977a3 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_flush.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_flush.py @@ -38,6 +38,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/StorageTargets_Flush.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/StorageTargets_Flush.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_get.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_get.py index 8ef0c1320c1f..61cfbb5aa39b 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_get.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_get.py @@ -39,6 +39,6 @@ def main(): print(response) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/StorageTargets_Get.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/StorageTargets_Get.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_invalidate.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_invalidate.py index 2ef51d8fac03..cb330c73503f 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_invalidate.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_invalidate.py @@ -38,6 +38,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/StorageTargets_Invalidate.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/StorageTargets_Invalidate.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_list_by_cache.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_list_by_cache.py index 83a701d43e8a..1c28eb01a7ba 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_list_by_cache.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_list_by_cache.py @@ -39,6 +39,6 @@ def main(): print(item) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/StorageTargets_ListByCache.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/StorageTargets_ListByCache.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_restore_defaults.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_restore_defaults.py index cadb1d2fea95..f6637c12e56f 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_restore_defaults.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_restore_defaults.py @@ -38,6 +38,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/StorageTargets_RestoreDefaults.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/StorageTargets_RestoreDefaults.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_resume.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_resume.py index 8fcb113cf21a..429dd867ce3c 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_resume.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_resume.py @@ -38,6 +38,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/StorageTargets_Resume.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/StorageTargets_Resume.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_suspend.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_suspend.py index 04369d3b3330..5dc813fd0c7e 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_suspend.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/storage_targets_suspend.py @@ -38,6 +38,6 @@ def main(): ).result() -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/StorageTargets_Suspend.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/StorageTargets_Suspend.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_samples/usage_models_list.py b/sdk/storage/azure-mgmt-storagecache/generated_samples/usage_models_list.py index e0000a658e73..c68e4cd5fec8 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_samples/usage_models_list.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_samples/usage_models_list.py @@ -36,6 +36,6 @@ def main(): print(item) -# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/stable/2025-07-01/examples/UsageModels_List.json +# x-ms-original-file: specification/storagecache/resource-manager/Microsoft.StorageCache/StorageCache/stable/2026-01-01/examples/UsageModels_List.json if __name__ == "__main__": main() diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management.py index 9dc9beb7d792..5d108a570c56 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management.py @@ -22,7 +22,7 @@ def setup_method(self, method): @recorded_by_proxy def test_check_aml_fs_subnets(self, resource_group): response = self.client.check_aml_fs_subnets( - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself @@ -32,7 +32,7 @@ def test_check_aml_fs_subnets(self, resource_group): @recorded_by_proxy def test_get_required_aml_fs_subnets_size(self, resource_group): response = self.client.get_required_aml_fs_subnets_size( - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_aml_filesystems_operations.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_aml_filesystems_operations.py index 2f0b29cdfeda..6c0fd0b22789 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_aml_filesystems_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_aml_filesystems_operations.py @@ -22,7 +22,7 @@ def setup_method(self, method): @recorded_by_proxy def test_aml_filesystems_list(self, resource_group): response = self.client.aml_filesystems.list( - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r for r in response] # please add some check logic here by yourself @@ -33,7 +33,7 @@ def test_aml_filesystems_list(self, resource_group): def test_aml_filesystems_list_by_resource_group(self, resource_group): response = self.client.aml_filesystems.list_by_resource_group( resource_group_name=resource_group.name, - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r for r in response] # please add some check logic here by yourself @@ -45,7 +45,7 @@ def test_aml_filesystems_begin_delete(self, resource_group): response = self.client.aml_filesystems.begin_delete( resource_group_name=resource_group.name, aml_filesystem_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -57,7 +57,7 @@ def test_aml_filesystems_get(self, resource_group): response = self.client.aml_filesystems.get( resource_group_name=resource_group.name, aml_filesystem_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself @@ -81,6 +81,8 @@ def test_aml_filesystems_begin_create_or_update(self, resource_group): "mgsAddress": "str", "mountCommand": "str", }, + "clusterUuid": "str", + "currentStorageCapacityTiB": 0.0, "encryptionSettings": {"keyEncryptionKey": {"keyUrl": "str", "sourceVault": {"id": "str"}}}, "filesystemSubnet": "str", "health": {"state": "str", "statusCode": "str", "statusDescription": "str"}, @@ -137,7 +139,7 @@ def test_aml_filesystems_begin_create_or_update(self, resource_group): "type": "str", "zones": ["str"], }, - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -161,7 +163,7 @@ def test_aml_filesystems_begin_update(self, resource_group): }, "tags": {"str": "str"}, }, - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -173,7 +175,7 @@ def test_aml_filesystems_archive(self, resource_group): response = self.client.aml_filesystems.archive( resource_group_name=resource_group.name, aml_filesystem_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself @@ -185,7 +187,7 @@ def test_aml_filesystems_cancel_archive(self, resource_group): response = self.client.aml_filesystems.cancel_archive( resource_group_name=resource_group.name, aml_filesystem_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_aml_filesystems_operations_async.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_aml_filesystems_operations_async.py index cf4c5bc90d6d..5418c3c08421 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_aml_filesystems_operations_async.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_aml_filesystems_operations_async.py @@ -23,7 +23,7 @@ def setup_method(self, method): @recorded_by_proxy_async async def test_aml_filesystems_list(self, resource_group): response = self.client.aml_filesystems.list( - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r async for r in response] # please add some check logic here by yourself @@ -34,7 +34,7 @@ async def test_aml_filesystems_list(self, resource_group): async def test_aml_filesystems_list_by_resource_group(self, resource_group): response = self.client.aml_filesystems.list_by_resource_group( resource_group_name=resource_group.name, - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r async for r in response] # please add some check logic here by yourself @@ -47,7 +47,7 @@ async def test_aml_filesystems_begin_delete(self, resource_group): await self.client.aml_filesystems.begin_delete( resource_group_name=resource_group.name, aml_filesystem_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -60,7 +60,7 @@ async def test_aml_filesystems_get(self, resource_group): response = await self.client.aml_filesystems.get( resource_group_name=resource_group.name, aml_filesystem_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself @@ -85,6 +85,8 @@ async def test_aml_filesystems_begin_create_or_update(self, resource_group): "mgsAddress": "str", "mountCommand": "str", }, + "clusterUuid": "str", + "currentStorageCapacityTiB": 0.0, "encryptionSettings": {"keyEncryptionKey": {"keyUrl": "str", "sourceVault": {"id": "str"}}}, "filesystemSubnet": "str", "health": {"state": "str", "statusCode": "str", "statusDescription": "str"}, @@ -141,7 +143,7 @@ async def test_aml_filesystems_begin_create_or_update(self, resource_group): "type": "str", "zones": ["str"], }, - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -167,7 +169,7 @@ async def test_aml_filesystems_begin_update(self, resource_group): }, "tags": {"str": "str"}, }, - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -180,7 +182,7 @@ async def test_aml_filesystems_archive(self, resource_group): response = await self.client.aml_filesystems.archive( resource_group_name=resource_group.name, aml_filesystem_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself @@ -192,7 +194,7 @@ async def test_aml_filesystems_cancel_archive(self, resource_group): response = await self.client.aml_filesystems.cancel_archive( resource_group_name=resource_group.name, aml_filesystem_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_asc_operations_operations.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_asc_operations_operations.py index 7cc43004e0d1..497afe3dbbc1 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_asc_operations_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_asc_operations_operations.py @@ -24,7 +24,7 @@ def test_asc_operations_get(self, resource_group): response = self.client.asc_operations.get( location="str", operation_id="str", - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_asc_operations_operations_async.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_asc_operations_operations_async.py index 51b23f139045..05a136f3434f 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_asc_operations_operations_async.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_asc_operations_operations_async.py @@ -25,7 +25,7 @@ async def test_asc_operations_get(self, resource_group): response = await self.client.asc_operations.get( location="str", operation_id="str", - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_asc_usages_operations.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_asc_usages_operations.py index 06860bfc149e..43a593a4310a 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_asc_usages_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_asc_usages_operations.py @@ -23,7 +23,7 @@ def setup_method(self, method): def test_asc_usages_list(self, resource_group): response = self.client.asc_usages.list( location="str", - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r for r in response] # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_asc_usages_operations_async.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_asc_usages_operations_async.py index ed3d509a6a26..b4d7288c800c 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_asc_usages_operations_async.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_asc_usages_operations_async.py @@ -24,7 +24,7 @@ def setup_method(self, method): async def test_asc_usages_list(self, resource_group): response = self.client.asc_usages.list( location="str", - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r async for r in response] # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_async.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_async.py index 1fd2cb2d3586..ac2258d70a6c 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_async.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_async.py @@ -23,7 +23,7 @@ def setup_method(self, method): @recorded_by_proxy_async async def test_check_aml_fs_subnets(self, resource_group): response = await self.client.check_aml_fs_subnets( - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself @@ -33,7 +33,7 @@ async def test_check_aml_fs_subnets(self, resource_group): @recorded_by_proxy_async async def test_get_required_aml_fs_subnets_size(self, resource_group): response = await self.client.get_required_aml_fs_subnets_size( - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_auto_export_jobs_operations.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_auto_export_jobs_operations.py index 7a0f17f50569..c9b7d4f8ee96 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_auto_export_jobs_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_auto_export_jobs_operations.py @@ -25,7 +25,7 @@ def test_auto_export_jobs_begin_delete(self, resource_group): resource_group_name=resource_group.name, aml_filesystem_name="str", auto_export_job_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -38,7 +38,7 @@ def test_auto_export_jobs_get(self, resource_group): resource_group_name=resource_group.name, aml_filesystem_name="str", auto_export_job_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself @@ -84,7 +84,7 @@ def test_auto_export_jobs_begin_create_or_update(self, resource_group): "totalMiBExported": 0, "type": "str", }, - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -98,7 +98,7 @@ def test_auto_export_jobs_begin_update(self, resource_group): aml_filesystem_name="str", auto_export_job_name="str", auto_export_job={"adminStatus": "str", "tags": {"str": "str"}}, - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -110,7 +110,7 @@ def test_auto_export_jobs_list_by_aml_filesystem(self, resource_group): response = self.client.auto_export_jobs.list_by_aml_filesystem( resource_group_name=resource_group.name, aml_filesystem_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r for r in response] # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_auto_export_jobs_operations_async.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_auto_export_jobs_operations_async.py index e14a77ce69b9..8790851d980f 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_auto_export_jobs_operations_async.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_auto_export_jobs_operations_async.py @@ -27,7 +27,7 @@ async def test_auto_export_jobs_begin_delete(self, resource_group): resource_group_name=resource_group.name, aml_filesystem_name="str", auto_export_job_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -41,7 +41,7 @@ async def test_auto_export_jobs_get(self, resource_group): resource_group_name=resource_group.name, aml_filesystem_name="str", auto_export_job_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself @@ -88,7 +88,7 @@ async def test_auto_export_jobs_begin_create_or_update(self, resource_group): "totalMiBExported": 0, "type": "str", }, - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -104,7 +104,7 @@ async def test_auto_export_jobs_begin_update(self, resource_group): aml_filesystem_name="str", auto_export_job_name="str", auto_export_job={"adminStatus": "str", "tags": {"str": "str"}}, - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -117,7 +117,7 @@ async def test_auto_export_jobs_list_by_aml_filesystem(self, resource_group): response = self.client.auto_export_jobs.list_by_aml_filesystem( resource_group_name=resource_group.name, aml_filesystem_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r async for r in response] # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_auto_import_jobs_operations.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_auto_import_jobs_operations.py index ccfdbc95801d..8825d5b7cea4 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_auto_import_jobs_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_auto_import_jobs_operations.py @@ -25,7 +25,7 @@ def test_auto_import_jobs_begin_delete(self, resource_group): resource_group_name=resource_group.name, aml_filesystem_name="str", auto_import_job_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -38,7 +38,7 @@ def test_auto_import_jobs_get(self, resource_group): resource_group_name=resource_group.name, aml_filesystem_name="str", auto_import_job_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself @@ -106,7 +106,7 @@ def test_auto_import_jobs_begin_create_or_update(self, resource_group): "totalErrors": 0, "type": "str", }, - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -120,7 +120,7 @@ def test_auto_import_jobs_begin_update(self, resource_group): aml_filesystem_name="str", auto_import_job_name="str", auto_import_job={"adminStatus": "str", "tags": {"str": "str"}}, - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -132,7 +132,7 @@ def test_auto_import_jobs_list_by_aml_filesystem(self, resource_group): response = self.client.auto_import_jobs.list_by_aml_filesystem( resource_group_name=resource_group.name, aml_filesystem_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r for r in response] # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_auto_import_jobs_operations_async.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_auto_import_jobs_operations_async.py index 8297fe251a16..ee131c882a34 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_auto_import_jobs_operations_async.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_auto_import_jobs_operations_async.py @@ -27,7 +27,7 @@ async def test_auto_import_jobs_begin_delete(self, resource_group): resource_group_name=resource_group.name, aml_filesystem_name="str", auto_import_job_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -41,7 +41,7 @@ async def test_auto_import_jobs_get(self, resource_group): resource_group_name=resource_group.name, aml_filesystem_name="str", auto_import_job_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself @@ -110,7 +110,7 @@ async def test_auto_import_jobs_begin_create_or_update(self, resource_group): "totalErrors": 0, "type": "str", }, - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -126,7 +126,7 @@ async def test_auto_import_jobs_begin_update(self, resource_group): aml_filesystem_name="str", auto_import_job_name="str", auto_import_job={"adminStatus": "str", "tags": {"str": "str"}}, - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -139,7 +139,7 @@ async def test_auto_import_jobs_list_by_aml_filesystem(self, resource_group): response = self.client.auto_import_jobs.list_by_aml_filesystem( resource_group_name=resource_group.name, aml_filesystem_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r async for r in response] # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_caches_operations.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_caches_operations.py index a0a7570aaac2..0ae3db6a7822 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_caches_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_caches_operations.py @@ -22,7 +22,7 @@ def setup_method(self, method): @recorded_by_proxy def test_caches_list(self, resource_group): response = self.client.caches.list( - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r for r in response] # please add some check logic here by yourself @@ -33,7 +33,7 @@ def test_caches_list(self, resource_group): def test_caches_list_by_resource_group(self, resource_group): response = self.client.caches.list_by_resource_group( resource_group_name=resource_group.name, - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r for r in response] # please add some check logic here by yourself @@ -45,7 +45,7 @@ def test_caches_begin_delete(self, resource_group): response = self.client.caches.begin_delete( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -57,7 +57,7 @@ def test_caches_get(self, resource_group): response = self.client.caches.get( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself @@ -176,7 +176,7 @@ def test_caches_begin_create_or_update(self, resource_group): }, "zones": ["str"], }, - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -188,7 +188,7 @@ def test_caches_begin_update(self, resource_group): response = self.client.caches.begin_update( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -200,7 +200,7 @@ def test_caches_begin_debug_info(self, resource_group): response = self.client.caches.begin_debug_info( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -212,7 +212,7 @@ def test_caches_begin_flush(self, resource_group): response = self.client.caches.begin_flush( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -224,7 +224,7 @@ def test_caches_begin_start(self, resource_group): response = self.client.caches.begin_start( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -236,7 +236,7 @@ def test_caches_begin_stop(self, resource_group): response = self.client.caches.begin_stop( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -248,7 +248,7 @@ def test_caches_begin_start_priming_job(self, resource_group): response = self.client.caches.begin_start_priming_job( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -260,7 +260,7 @@ def test_caches_begin_stop_priming_job(self, resource_group): response = self.client.caches.begin_stop_priming_job( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -272,7 +272,7 @@ def test_caches_begin_pause_priming_job(self, resource_group): response = self.client.caches.begin_pause_priming_job( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -284,7 +284,7 @@ def test_caches_begin_resume_priming_job(self, resource_group): response = self.client.caches.begin_resume_priming_job( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -296,7 +296,7 @@ def test_caches_begin_upgrade_firmware(self, resource_group): response = self.client.caches.begin_upgrade_firmware( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -308,7 +308,7 @@ def test_caches_begin_space_allocation(self, resource_group): response = self.client.caches.begin_space_allocation( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_caches_operations_async.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_caches_operations_async.py index 5742d7d56034..d83fb63c36fc 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_caches_operations_async.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_caches_operations_async.py @@ -23,7 +23,7 @@ def setup_method(self, method): @recorded_by_proxy_async async def test_caches_list(self, resource_group): response = self.client.caches.list( - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r async for r in response] # please add some check logic here by yourself @@ -34,7 +34,7 @@ async def test_caches_list(self, resource_group): async def test_caches_list_by_resource_group(self, resource_group): response = self.client.caches.list_by_resource_group( resource_group_name=resource_group.name, - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r async for r in response] # please add some check logic here by yourself @@ -47,7 +47,7 @@ async def test_caches_begin_delete(self, resource_group): await self.client.caches.begin_delete( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -60,7 +60,7 @@ async def test_caches_get(self, resource_group): response = await self.client.caches.get( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself @@ -180,7 +180,7 @@ async def test_caches_begin_create_or_update(self, resource_group): }, "zones": ["str"], }, - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -194,7 +194,7 @@ async def test_caches_begin_update(self, resource_group): await self.client.caches.begin_update( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -208,7 +208,7 @@ async def test_caches_begin_debug_info(self, resource_group): await self.client.caches.begin_debug_info( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -222,7 +222,7 @@ async def test_caches_begin_flush(self, resource_group): await self.client.caches.begin_flush( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -236,7 +236,7 @@ async def test_caches_begin_start(self, resource_group): await self.client.caches.begin_start( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -250,7 +250,7 @@ async def test_caches_begin_stop(self, resource_group): await self.client.caches.begin_stop( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -264,7 +264,7 @@ async def test_caches_begin_start_priming_job(self, resource_group): await self.client.caches.begin_start_priming_job( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -278,7 +278,7 @@ async def test_caches_begin_stop_priming_job(self, resource_group): await self.client.caches.begin_stop_priming_job( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -292,7 +292,7 @@ async def test_caches_begin_pause_priming_job(self, resource_group): await self.client.caches.begin_pause_priming_job( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -306,7 +306,7 @@ async def test_caches_begin_resume_priming_job(self, resource_group): await self.client.caches.begin_resume_priming_job( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -320,7 +320,7 @@ async def test_caches_begin_upgrade_firmware(self, resource_group): await self.client.caches.begin_upgrade_firmware( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -334,7 +334,7 @@ async def test_caches_begin_space_allocation(self, resource_group): await self.client.caches.begin_space_allocation( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_expansion_jobs_operations.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_expansion_jobs_operations.py new file mode 100644 index 000000000000..56a72f18093d --- /dev/null +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_expansion_jobs_operations.py @@ -0,0 +1,107 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- +import pytest +from azure.mgmt.storagecache import StorageCacheManagementClient + +from devtools_testutils import AzureMgmtRecordedTestCase, RandomNameResourceGroupPreparer, recorded_by_proxy + +AZURE_LOCATION = "eastus" + + +@pytest.mark.skip("you may need to update the auto-generated test case before run it") +class TestStorageCacheManagementExpansionJobsOperations(AzureMgmtRecordedTestCase): + def setup_method(self, method): + self.client = self.create_mgmt_client(StorageCacheManagementClient) + + @RandomNameResourceGroupPreparer(location=AZURE_LOCATION) + @recorded_by_proxy + def test_expansion_jobs_begin_delete(self, resource_group): + response = self.client.expansion_jobs.begin_delete( + resource_group_name=resource_group.name, + aml_filesystem_name="str", + expansion_job_name="str", + api_version="2026-01-01", + ).result() # call '.result()' to poll until service return final result + + # please add some check logic here by yourself + # ... + + @RandomNameResourceGroupPreparer(location=AZURE_LOCATION) + @recorded_by_proxy + def test_expansion_jobs_get(self, resource_group): + response = self.client.expansion_jobs.get( + resource_group_name=resource_group.name, + aml_filesystem_name="str", + expansion_job_name="str", + api_version="2026-01-01", + ) + + # please add some check logic here by yourself + # ... + + @RandomNameResourceGroupPreparer(location=AZURE_LOCATION) + @recorded_by_proxy + def test_expansion_jobs_begin_create_or_update(self, resource_group): + response = self.client.expansion_jobs.begin_create_or_update( + resource_group_name=resource_group.name, + aml_filesystem_name="str", + expansion_job_name="str", + expansion_job={ + "location": "str", + "completionTimeUTC": "2020-02-20 00:00:00", + "id": "str", + "name": "str", + "newStorageCapacityTiB": 0.0, + "percentComplete": 0.0, + "provisioningState": "str", + "startTimeUTC": "2020-02-20 00:00:00", + "state": "str", + "statusCode": "str", + "statusMessage": "str", + "systemData": { + "createdAt": "2020-02-20 00:00:00", + "createdBy": "str", + "createdByType": "str", + "lastModifiedAt": "2020-02-20 00:00:00", + "lastModifiedBy": "str", + "lastModifiedByType": "str", + }, + "tags": {"str": "str"}, + "type": "str", + }, + api_version="2026-01-01", + ).result() # call '.result()' to poll until service return final result + + # please add some check logic here by yourself + # ... + + @RandomNameResourceGroupPreparer(location=AZURE_LOCATION) + @recorded_by_proxy + def test_expansion_jobs_begin_update(self, resource_group): + response = self.client.expansion_jobs.begin_update( + resource_group_name=resource_group.name, + aml_filesystem_name="str", + expansion_job_name="str", + expansion_job={"tags": {"str": "str"}}, + api_version="2026-01-01", + ).result() # call '.result()' to poll until service return final result + + # please add some check logic here by yourself + # ... + + @RandomNameResourceGroupPreparer(location=AZURE_LOCATION) + @recorded_by_proxy + def test_expansion_jobs_list_by_aml_filesystem(self, resource_group): + response = self.client.expansion_jobs.list_by_aml_filesystem( + resource_group_name=resource_group.name, + aml_filesystem_name="str", + api_version="2026-01-01", + ) + result = [r for r in response] + # please add some check logic here by yourself + # ... diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_expansion_jobs_operations_async.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_expansion_jobs_operations_async.py new file mode 100644 index 000000000000..0c8e3171dd8c --- /dev/null +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_expansion_jobs_operations_async.py @@ -0,0 +1,114 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for license information. +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is regenerated. +# -------------------------------------------------------------------------- +import pytest +from azure.mgmt.storagecache.aio import StorageCacheManagementClient + +from devtools_testutils import AzureMgmtRecordedTestCase, RandomNameResourceGroupPreparer +from devtools_testutils.aio import recorded_by_proxy_async + +AZURE_LOCATION = "eastus" + + +@pytest.mark.skip("you may need to update the auto-generated test case before run it") +class TestStorageCacheManagementExpansionJobsOperationsAsync(AzureMgmtRecordedTestCase): + def setup_method(self, method): + self.client = self.create_mgmt_client(StorageCacheManagementClient, is_async=True) + + @RandomNameResourceGroupPreparer(location=AZURE_LOCATION) + @recorded_by_proxy_async + async def test_expansion_jobs_begin_delete(self, resource_group): + response = await ( + await self.client.expansion_jobs.begin_delete( + resource_group_name=resource_group.name, + aml_filesystem_name="str", + expansion_job_name="str", + api_version="2026-01-01", + ) + ).result() # call '.result()' to poll until service return final result + + # please add some check logic here by yourself + # ... + + @RandomNameResourceGroupPreparer(location=AZURE_LOCATION) + @recorded_by_proxy_async + async def test_expansion_jobs_get(self, resource_group): + response = await self.client.expansion_jobs.get( + resource_group_name=resource_group.name, + aml_filesystem_name="str", + expansion_job_name="str", + api_version="2026-01-01", + ) + + # please add some check logic here by yourself + # ... + + @RandomNameResourceGroupPreparer(location=AZURE_LOCATION) + @recorded_by_proxy_async + async def test_expansion_jobs_begin_create_or_update(self, resource_group): + response = await ( + await self.client.expansion_jobs.begin_create_or_update( + resource_group_name=resource_group.name, + aml_filesystem_name="str", + expansion_job_name="str", + expansion_job={ + "location": "str", + "completionTimeUTC": "2020-02-20 00:00:00", + "id": "str", + "name": "str", + "newStorageCapacityTiB": 0.0, + "percentComplete": 0.0, + "provisioningState": "str", + "startTimeUTC": "2020-02-20 00:00:00", + "state": "str", + "statusCode": "str", + "statusMessage": "str", + "systemData": { + "createdAt": "2020-02-20 00:00:00", + "createdBy": "str", + "createdByType": "str", + "lastModifiedAt": "2020-02-20 00:00:00", + "lastModifiedBy": "str", + "lastModifiedByType": "str", + }, + "tags": {"str": "str"}, + "type": "str", + }, + api_version="2026-01-01", + ) + ).result() # call '.result()' to poll until service return final result + + # please add some check logic here by yourself + # ... + + @RandomNameResourceGroupPreparer(location=AZURE_LOCATION) + @recorded_by_proxy_async + async def test_expansion_jobs_begin_update(self, resource_group): + response = await ( + await self.client.expansion_jobs.begin_update( + resource_group_name=resource_group.name, + aml_filesystem_name="str", + expansion_job_name="str", + expansion_job={"tags": {"str": "str"}}, + api_version="2026-01-01", + ) + ).result() # call '.result()' to poll until service return final result + + # please add some check logic here by yourself + # ... + + @RandomNameResourceGroupPreparer(location=AZURE_LOCATION) + @recorded_by_proxy_async + async def test_expansion_jobs_list_by_aml_filesystem(self, resource_group): + response = self.client.expansion_jobs.list_by_aml_filesystem( + resource_group_name=resource_group.name, + aml_filesystem_name="str", + api_version="2026-01-01", + ) + result = [r async for r in response] + # please add some check logic here by yourself + # ... diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_import_jobs_operations.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_import_jobs_operations.py index bdfc8449c327..90b419895c6e 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_import_jobs_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_import_jobs_operations.py @@ -25,7 +25,7 @@ def test_import_jobs_begin_delete(self, resource_group): resource_group_name=resource_group.name, aml_filesystem_name="str", import_job_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -38,7 +38,7 @@ def test_import_jobs_get(self, resource_group): resource_group_name=resource_group.name, aml_filesystem_name="str", import_job_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself @@ -87,7 +87,7 @@ def test_import_jobs_begin_create_or_update(self, resource_group): "totalErrors": 0, "type": "str", }, - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -101,7 +101,7 @@ def test_import_jobs_begin_update(self, resource_group): aml_filesystem_name="str", import_job_name="str", import_job={"adminStatus": "str", "tags": {"str": "str"}}, - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -113,7 +113,7 @@ def test_import_jobs_list_by_aml_filesystem(self, resource_group): response = self.client.import_jobs.list_by_aml_filesystem( resource_group_name=resource_group.name, aml_filesystem_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r for r in response] # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_import_jobs_operations_async.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_import_jobs_operations_async.py index b803f7209d8f..be3287fd1510 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_import_jobs_operations_async.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_import_jobs_operations_async.py @@ -27,7 +27,7 @@ async def test_import_jobs_begin_delete(self, resource_group): resource_group_name=resource_group.name, aml_filesystem_name="str", import_job_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -41,7 +41,7 @@ async def test_import_jobs_get(self, resource_group): resource_group_name=resource_group.name, aml_filesystem_name="str", import_job_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself @@ -91,7 +91,7 @@ async def test_import_jobs_begin_create_or_update(self, resource_group): "totalErrors": 0, "type": "str", }, - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -107,7 +107,7 @@ async def test_import_jobs_begin_update(self, resource_group): aml_filesystem_name="str", import_job_name="str", import_job={"adminStatus": "str", "tags": {"str": "str"}}, - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -120,7 +120,7 @@ async def test_import_jobs_list_by_aml_filesystem(self, resource_group): response = self.client.import_jobs.list_by_aml_filesystem( resource_group_name=resource_group.name, aml_filesystem_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r async for r in response] # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_operations.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_operations.py index eb88196f30a6..13b0307311cc 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_operations.py @@ -22,7 +22,7 @@ def setup_method(self, method): @recorded_by_proxy def test_operations_list(self, resource_group): response = self.client.operations.list( - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r for r in response] # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_operations_async.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_operations_async.py index 7f87468708b9..cee31933e482 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_operations_async.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_operations_async.py @@ -23,7 +23,7 @@ def setup_method(self, method): @recorded_by_proxy_async async def test_operations_list(self, resource_group): response = self.client.operations.list( - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r async for r in response] # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_skus_operations.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_skus_operations.py index 63c08cb45498..8e46a88813cd 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_skus_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_skus_operations.py @@ -22,7 +22,7 @@ def setup_method(self, method): @recorded_by_proxy def test_skus_list(self, resource_group): response = self.client.skus.list( - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r for r in response] # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_skus_operations_async.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_skus_operations_async.py index 61f7eca47b5f..24257107bbfb 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_skus_operations_async.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_skus_operations_async.py @@ -23,7 +23,7 @@ def setup_method(self, method): @recorded_by_proxy_async async def test_skus_list(self, resource_group): response = self.client.skus.list( - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r async for r in response] # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_storage_target_operations.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_storage_target_operations.py index c541084b75dd..f97e169ef1b9 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_storage_target_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_storage_target_operations.py @@ -25,7 +25,7 @@ def test_storage_target_begin_flush(self, resource_group): resource_group_name=resource_group.name, cache_name="str", storage_target_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -38,7 +38,7 @@ def test_storage_target_begin_suspend(self, resource_group): resource_group_name=resource_group.name, cache_name="str", storage_target_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -51,7 +51,7 @@ def test_storage_target_begin_resume(self, resource_group): resource_group_name=resource_group.name, cache_name="str", storage_target_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -64,7 +64,7 @@ def test_storage_target_begin_invalidate(self, resource_group): resource_group_name=resource_group.name, cache_name="str", storage_target_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_storage_target_operations_async.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_storage_target_operations_async.py index a81d3ed0f6c7..eaf806c63e94 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_storage_target_operations_async.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_storage_target_operations_async.py @@ -27,7 +27,7 @@ async def test_storage_target_begin_flush(self, resource_group): resource_group_name=resource_group.name, cache_name="str", storage_target_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -42,7 +42,7 @@ async def test_storage_target_begin_suspend(self, resource_group): resource_group_name=resource_group.name, cache_name="str", storage_target_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -57,7 +57,7 @@ async def test_storage_target_begin_resume(self, resource_group): resource_group_name=resource_group.name, cache_name="str", storage_target_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -72,7 +72,7 @@ async def test_storage_target_begin_invalidate(self, resource_group): resource_group_name=resource_group.name, cache_name="str", storage_target_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_storage_targets_operations.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_storage_targets_operations.py index 373a633ec3d9..094a16f9a45f 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_storage_targets_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_storage_targets_operations.py @@ -25,7 +25,7 @@ def test_storage_targets_begin_dns_refresh(self, resource_group): resource_group_name=resource_group.name, cache_name="str", storage_target_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -37,7 +37,7 @@ def test_storage_targets_list_by_cache(self, resource_group): response = self.client.storage_targets.list_by_cache( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r for r in response] # please add some check logic here by yourself @@ -50,7 +50,7 @@ def test_storage_targets_begin_delete(self, resource_group): resource_group_name=resource_group.name, cache_name="str", storage_target_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -63,7 +63,7 @@ def test_storage_targets_get(self, resource_group): resource_group_name=resource_group.name, cache_name="str", storage_target_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself @@ -101,7 +101,7 @@ def test_storage_targets_begin_create_or_update(self, resource_group): "type": "str", "unknown": {"attributes": {"str": "str"}}, }, - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself @@ -114,7 +114,7 @@ def test_storage_targets_begin_restore_defaults(self, resource_group): resource_group_name=resource_group.name, cache_name="str", storage_target_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ).result() # call '.result()' to poll until service return final result # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_storage_targets_operations_async.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_storage_targets_operations_async.py index fb13f46516af..54dc8eb48fa9 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_storage_targets_operations_async.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_storage_targets_operations_async.py @@ -27,7 +27,7 @@ async def test_storage_targets_begin_dns_refresh(self, resource_group): resource_group_name=resource_group.name, cache_name="str", storage_target_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -40,7 +40,7 @@ async def test_storage_targets_list_by_cache(self, resource_group): response = self.client.storage_targets.list_by_cache( resource_group_name=resource_group.name, cache_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r async for r in response] # please add some check logic here by yourself @@ -54,7 +54,7 @@ async def test_storage_targets_begin_delete(self, resource_group): resource_group_name=resource_group.name, cache_name="str", storage_target_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -68,7 +68,7 @@ async def test_storage_targets_get(self, resource_group): resource_group_name=resource_group.name, cache_name="str", storage_target_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) # please add some check logic here by yourself @@ -107,7 +107,7 @@ async def test_storage_targets_begin_create_or_update(self, resource_group): "type": "str", "unknown": {"attributes": {"str": "str"}}, }, - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result @@ -122,7 +122,7 @@ async def test_storage_targets_begin_restore_defaults(self, resource_group): resource_group_name=resource_group.name, cache_name="str", storage_target_name="str", - api_version="2025-07-01", + api_version="2026-01-01", ) ).result() # call '.result()' to poll until service return final result diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_usage_models_operations.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_usage_models_operations.py index c9e91f3e90bf..bf400a1f095e 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_usage_models_operations.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_usage_models_operations.py @@ -22,7 +22,7 @@ def setup_method(self, method): @recorded_by_proxy def test_usage_models_list(self, resource_group): response = self.client.usage_models.list( - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r for r in response] # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_usage_models_operations_async.py b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_usage_models_operations_async.py index 244265588258..8dac90545e4b 100644 --- a/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_usage_models_operations_async.py +++ b/sdk/storage/azure-mgmt-storagecache/generated_tests/test_storage_cache_management_usage_models_operations_async.py @@ -23,7 +23,7 @@ def setup_method(self, method): @recorded_by_proxy_async async def test_usage_models_list(self, resource_group): response = self.client.usage_models.list( - api_version="2025-07-01", + api_version="2026-01-01", ) result = [r async for r in response] # please add some check logic here by yourself diff --git a/sdk/storage/azure-mgmt-storagecache/pyproject.toml b/sdk/storage/azure-mgmt-storagecache/pyproject.toml index 899eb20c5429..e346f3a26f36 100644 --- a/sdk/storage/azure-mgmt-storagecache/pyproject.toml +++ b/sdk/storage/azure-mgmt-storagecache/pyproject.toml @@ -29,8 +29,8 @@ keywords = [ "azure sdk", ] dependencies = [ - "msrest>=0.7.1", - "azure-mgmt-core>=1.5.0", + "isodate>=0.6.1", + "azure-mgmt-core>=1.6.0", "typing-extensions>=4.6.0", ] dynamic = [ @@ -54,8 +54,8 @@ content-type = "text/markdown" [tool.setuptools.packages.find] exclude = [ "tests*", - "samples*", "generated_tests*", + "samples*", "generated_samples*", "doc*", "azure",